var/home/core/zuul-output/0000755000175000017500000000000015114202406014520 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114216446015475 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005051563115114216440017701 0ustar rootrootDec 04 03:54:44 crc systemd[1]: Starting Kubernetes Kubelet... Dec 04 03:54:44 crc restorecon[4583]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:44 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 03:54:45 crc restorecon[4583]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 03:54:45 crc restorecon[4583]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 04 03:54:45 crc kubenswrapper[4806]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 03:54:45 crc kubenswrapper[4806]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 04 03:54:45 crc kubenswrapper[4806]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 03:54:45 crc kubenswrapper[4806]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 03:54:45 crc kubenswrapper[4806]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 04 03:54:45 crc kubenswrapper[4806]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.251413 4806 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255543 4806 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255561 4806 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255565 4806 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255570 4806 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255574 4806 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255579 4806 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255583 4806 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255588 4806 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255594 4806 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255599 4806 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255606 4806 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255610 4806 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255613 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255617 4806 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255621 4806 feature_gate.go:330] unrecognized feature gate: Example Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255624 4806 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255629 4806 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255633 4806 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255639 4806 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255644 4806 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255649 4806 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255658 4806 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255663 4806 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255668 4806 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255673 4806 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255678 4806 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255683 4806 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255688 4806 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255692 4806 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255697 4806 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255701 4806 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255705 4806 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255710 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255716 4806 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255720 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255724 4806 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255728 4806 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255732 4806 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255736 4806 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255740 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255744 4806 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255747 4806 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255751 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255755 4806 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255758 4806 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255762 4806 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255768 4806 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255771 4806 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255775 4806 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255779 4806 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255783 4806 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255787 4806 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255793 4806 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255796 4806 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255800 4806 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255803 4806 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255807 4806 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255813 4806 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255817 4806 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255820 4806 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255823 4806 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255827 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255831 4806 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255834 4806 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255837 4806 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255841 4806 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255844 4806 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255848 4806 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255851 4806 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255855 4806 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.255860 4806 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.255970 4806 flags.go:64] FLAG: --address="0.0.0.0" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.255982 4806 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.255989 4806 flags.go:64] FLAG: --anonymous-auth="true" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.255995 4806 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256000 4806 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256005 4806 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256013 4806 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256019 4806 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256024 4806 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256028 4806 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256033 4806 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256038 4806 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256043 4806 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256047 4806 flags.go:64] FLAG: --cgroup-root="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256054 4806 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256058 4806 flags.go:64] FLAG: --client-ca-file="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256063 4806 flags.go:64] FLAG: --cloud-config="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256067 4806 flags.go:64] FLAG: --cloud-provider="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256071 4806 flags.go:64] FLAG: --cluster-dns="[]" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256076 4806 flags.go:64] FLAG: --cluster-domain="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256080 4806 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256084 4806 flags.go:64] FLAG: --config-dir="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256088 4806 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256095 4806 flags.go:64] FLAG: --container-log-max-files="5" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256101 4806 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256105 4806 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256109 4806 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256113 4806 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256118 4806 flags.go:64] FLAG: --contention-profiling="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256122 4806 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256126 4806 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256132 4806 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256137 4806 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256142 4806 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256146 4806 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256150 4806 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256155 4806 flags.go:64] FLAG: --enable-load-reader="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256160 4806 flags.go:64] FLAG: --enable-server="true" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256164 4806 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256169 4806 flags.go:64] FLAG: --event-burst="100" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256176 4806 flags.go:64] FLAG: --event-qps="50" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256180 4806 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256185 4806 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256190 4806 flags.go:64] FLAG: --eviction-hard="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256195 4806 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256200 4806 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256204 4806 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256208 4806 flags.go:64] FLAG: --eviction-soft="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256215 4806 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256219 4806 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256223 4806 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256227 4806 flags.go:64] FLAG: --experimental-mounter-path="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256231 4806 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256235 4806 flags.go:64] FLAG: --fail-swap-on="true" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256239 4806 flags.go:64] FLAG: --feature-gates="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256245 4806 flags.go:64] FLAG: --file-check-frequency="20s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256251 4806 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256255 4806 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256259 4806 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256264 4806 flags.go:64] FLAG: --healthz-port="10248" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256268 4806 flags.go:64] FLAG: --help="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256272 4806 flags.go:64] FLAG: --hostname-override="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256276 4806 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256280 4806 flags.go:64] FLAG: --http-check-frequency="20s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256285 4806 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256292 4806 flags.go:64] FLAG: --image-credential-provider-config="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256296 4806 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256300 4806 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256304 4806 flags.go:64] FLAG: --image-service-endpoint="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256308 4806 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256312 4806 flags.go:64] FLAG: --kube-api-burst="100" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256317 4806 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256322 4806 flags.go:64] FLAG: --kube-api-qps="50" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256327 4806 flags.go:64] FLAG: --kube-reserved="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256333 4806 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256337 4806 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256341 4806 flags.go:64] FLAG: --kubelet-cgroups="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256346 4806 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256350 4806 flags.go:64] FLAG: --lock-file="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256354 4806 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256358 4806 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256363 4806 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256372 4806 flags.go:64] FLAG: --log-json-split-stream="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256376 4806 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256380 4806 flags.go:64] FLAG: --log-text-split-stream="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256384 4806 flags.go:64] FLAG: --logging-format="text" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256388 4806 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256392 4806 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256396 4806 flags.go:64] FLAG: --manifest-url="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256400 4806 flags.go:64] FLAG: --manifest-url-header="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256406 4806 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256413 4806 flags.go:64] FLAG: --max-open-files="1000000" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256419 4806 flags.go:64] FLAG: --max-pods="110" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256424 4806 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256428 4806 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256433 4806 flags.go:64] FLAG: --memory-manager-policy="None" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256444 4806 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256448 4806 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256452 4806 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256459 4806 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256468 4806 flags.go:64] FLAG: --node-status-max-images="50" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256472 4806 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256477 4806 flags.go:64] FLAG: --oom-score-adj="-999" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256481 4806 flags.go:64] FLAG: --pod-cidr="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256485 4806 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256492 4806 flags.go:64] FLAG: --pod-manifest-path="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256499 4806 flags.go:64] FLAG: --pod-max-pids="-1" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256503 4806 flags.go:64] FLAG: --pods-per-core="0" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256508 4806 flags.go:64] FLAG: --port="10250" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256513 4806 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256518 4806 flags.go:64] FLAG: --provider-id="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256524 4806 flags.go:64] FLAG: --qos-reserved="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256529 4806 flags.go:64] FLAG: --read-only-port="10255" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256534 4806 flags.go:64] FLAG: --register-node="true" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256541 4806 flags.go:64] FLAG: --register-schedulable="true" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256546 4806 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256553 4806 flags.go:64] FLAG: --registry-burst="10" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256557 4806 flags.go:64] FLAG: --registry-qps="5" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256561 4806 flags.go:64] FLAG: --reserved-cpus="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256565 4806 flags.go:64] FLAG: --reserved-memory="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256571 4806 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256577 4806 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256581 4806 flags.go:64] FLAG: --rotate-certificates="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256586 4806 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256590 4806 flags.go:64] FLAG: --runonce="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256594 4806 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256599 4806 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256603 4806 flags.go:64] FLAG: --seccomp-default="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256607 4806 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256613 4806 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256619 4806 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256623 4806 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256628 4806 flags.go:64] FLAG: --storage-driver-password="root" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256632 4806 flags.go:64] FLAG: --storage-driver-secure="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256636 4806 flags.go:64] FLAG: --storage-driver-table="stats" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256640 4806 flags.go:64] FLAG: --storage-driver-user="root" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256644 4806 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256649 4806 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256653 4806 flags.go:64] FLAG: --system-cgroups="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256659 4806 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256666 4806 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256670 4806 flags.go:64] FLAG: --tls-cert-file="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256674 4806 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256680 4806 flags.go:64] FLAG: --tls-min-version="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256684 4806 flags.go:64] FLAG: --tls-private-key-file="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256691 4806 flags.go:64] FLAG: --topology-manager-policy="none" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256695 4806 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256699 4806 flags.go:64] FLAG: --topology-manager-scope="container" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256704 4806 flags.go:64] FLAG: --v="2" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256709 4806 flags.go:64] FLAG: --version="false" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256715 4806 flags.go:64] FLAG: --vmodule="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256720 4806 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.256724 4806 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.256947 4806 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.256955 4806 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257120 4806 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257129 4806 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257135 4806 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257140 4806 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257145 4806 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257152 4806 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257161 4806 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257166 4806 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257171 4806 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257175 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257179 4806 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257184 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257188 4806 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257192 4806 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257197 4806 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257203 4806 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257208 4806 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257213 4806 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257217 4806 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257222 4806 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257227 4806 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257232 4806 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257237 4806 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257241 4806 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257246 4806 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257251 4806 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257256 4806 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257260 4806 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257265 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257269 4806 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257274 4806 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257278 4806 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257284 4806 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257293 4806 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257303 4806 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257311 4806 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257315 4806 feature_gate.go:330] unrecognized feature gate: Example Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257320 4806 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257328 4806 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257333 4806 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257338 4806 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257342 4806 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257346 4806 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257350 4806 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257354 4806 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257359 4806 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257363 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257367 4806 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257371 4806 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257375 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257379 4806 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257386 4806 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257390 4806 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257394 4806 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257398 4806 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257401 4806 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257405 4806 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257409 4806 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257414 4806 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257418 4806 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257422 4806 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257426 4806 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257430 4806 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257433 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257437 4806 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257441 4806 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257444 4806 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257448 4806 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.257451 4806 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.257732 4806 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.264982 4806 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.265006 4806 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265091 4806 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265099 4806 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265105 4806 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265111 4806 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265117 4806 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265125 4806 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265132 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265138 4806 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265143 4806 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265149 4806 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265154 4806 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265159 4806 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265165 4806 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265170 4806 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265175 4806 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265181 4806 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265186 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265192 4806 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265197 4806 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265204 4806 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265211 4806 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265217 4806 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265222 4806 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265228 4806 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265242 4806 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265249 4806 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265254 4806 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265260 4806 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265266 4806 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265273 4806 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265279 4806 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265285 4806 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265291 4806 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265296 4806 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265302 4806 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265307 4806 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265313 4806 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265318 4806 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265323 4806 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265328 4806 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265333 4806 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265339 4806 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265344 4806 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265350 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265355 4806 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265360 4806 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265366 4806 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265371 4806 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265376 4806 feature_gate.go:330] unrecognized feature gate: Example Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265383 4806 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265389 4806 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265396 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265402 4806 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265407 4806 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265413 4806 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265418 4806 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265424 4806 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265430 4806 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265436 4806 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265442 4806 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265454 4806 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265461 4806 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265467 4806 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265472 4806 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265478 4806 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265484 4806 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265489 4806 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265495 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265502 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265507 4806 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265513 4806 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.265522 4806 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265670 4806 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265680 4806 feature_gate.go:330] unrecognized feature gate: Example Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265686 4806 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265696 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265704 4806 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265711 4806 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265717 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265723 4806 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265728 4806 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265734 4806 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265740 4806 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265746 4806 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265751 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265758 4806 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265763 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265769 4806 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265774 4806 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265780 4806 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265785 4806 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265790 4806 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265795 4806 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265801 4806 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265810 4806 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265817 4806 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265823 4806 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265829 4806 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265835 4806 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265840 4806 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265846 4806 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265851 4806 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265856 4806 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265862 4806 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265867 4806 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265873 4806 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265878 4806 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265883 4806 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265889 4806 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265894 4806 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265899 4806 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265906 4806 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265913 4806 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265919 4806 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265945 4806 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265951 4806 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265957 4806 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265962 4806 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265968 4806 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265973 4806 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265978 4806 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265984 4806 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265989 4806 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.265994 4806 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266000 4806 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266006 4806 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266012 4806 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266017 4806 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266022 4806 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266028 4806 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266033 4806 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266039 4806 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266044 4806 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266051 4806 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266057 4806 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266062 4806 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266067 4806 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266072 4806 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266079 4806 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266086 4806 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266092 4806 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266097 4806 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.266103 4806 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.266110 4806 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.266543 4806 server.go:940] "Client rotation is on, will bootstrap in background" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.269436 4806 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.269541 4806 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.270262 4806 server.go:997] "Starting client certificate rotation" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.270293 4806 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.270465 4806 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-27 16:07:28.68752558 +0000 UTC Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.270522 4806 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 564h12m43.417005991s for next certificate rotation Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.276322 4806 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.278670 4806 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.287458 4806 log.go:25] "Validated CRI v1 runtime API" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.307584 4806 log.go:25] "Validated CRI v1 image API" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.309757 4806 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.313064 4806 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-04-03-48-53-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.313116 4806 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.333029 4806 manager.go:217] Machine: {Timestamp:2025-12-04 03:54:45.331217155 +0000 UTC m=+0.189730163 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199480832 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:684cf60c-c6df-4023-948c-ef14ad787a60 BootID:5dbc708a-80bd-449f-9c89-8b2b0f364021 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:c3:cf:5b Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:c3:cf:5b Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:b8:74:c5 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:ea:73:d6 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:1f:7b:c4 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:a4:f3:f9 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:fe:7b:60:9f:73:f5 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:8e:cd:35:2c:db:02 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199480832 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.333375 4806 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.333588 4806 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.334036 4806 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.334410 4806 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.334475 4806 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.334799 4806 topology_manager.go:138] "Creating topology manager with none policy" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.334819 4806 container_manager_linux.go:303] "Creating device plugin manager" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.335163 4806 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.335230 4806 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.335645 4806 state_mem.go:36] "Initialized new in-memory state store" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.335780 4806 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.336691 4806 kubelet.go:418] "Attempting to sync node with API server" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.336730 4806 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.336771 4806 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.336794 4806 kubelet.go:324] "Adding apiserver pod source" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.336815 4806 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.339712 4806 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.340450 4806 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.340668 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.340675 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:54:45 crc kubenswrapper[4806]: E1204 03:54:45.340798 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Dec 04 03:54:45 crc kubenswrapper[4806]: E1204 03:54:45.340812 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.341776 4806 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.342958 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.343021 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.343076 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.343098 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.343129 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.343152 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.343173 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.343204 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.343226 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.343246 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.343293 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.343316 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.343670 4806 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.344685 4806 server.go:1280] "Started kubelet" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.345068 4806 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.345153 4806 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.345114 4806 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.346429 4806 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.346474 4806 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.346708 4806 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 17:21:48.299552152 +0000 UTC Dec 04 03:54:45 crc systemd[1]: Started Kubernetes Kubelet. Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.347426 4806 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.350981 4806 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.351019 4806 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 04 03:54:45 crc kubenswrapper[4806]: E1204 03:54:45.351160 4806 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.351236 4806 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.352227 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:54:45 crc kubenswrapper[4806]: E1204 03:54:45.352367 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.360744 4806 server.go:460] "Adding debug handlers to kubelet server" Dec 04 03:54:45 crc kubenswrapper[4806]: E1204 03:54:45.354017 4806 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.17:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187de6e1894ca784 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 03:54:45.344634756 +0000 UTC m=+0.203147734,LastTimestamp:2025-12-04 03:54:45.344634756 +0000 UTC m=+0.203147734,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 03:54:45 crc kubenswrapper[4806]: E1204 03:54:45.363143 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="200ms" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.363536 4806 factory.go:55] Registering systemd factory Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.363559 4806 factory.go:221] Registration of the systemd container factory successfully Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.364691 4806 factory.go:153] Registering CRI-O factory Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.364729 4806 factory.go:221] Registration of the crio container factory successfully Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.364830 4806 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.364861 4806 factory.go:103] Registering Raw factory Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.365118 4806 manager.go:1196] Started watching for new ooms in manager Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.366844 4806 manager.go:319] Starting recovery of all containers Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.382914 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383041 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383056 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383069 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383081 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383094 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383107 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383119 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383142 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383155 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383168 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383179 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383190 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383209 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383222 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383238 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383280 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383301 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383323 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383365 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383381 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383394 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383407 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383422 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383437 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383451 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383468 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383500 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383514 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383528 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383546 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383558 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383570 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383583 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383598 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.383610 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.388602 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.388625 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.388640 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.388653 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.388666 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.388678 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.388690 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.388702 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.388855 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.388913 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389017 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389050 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389092 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389164 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389197 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389227 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389266 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389309 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389384 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389423 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389454 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389561 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389589 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389659 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389722 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389780 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389846 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389894 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389973 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.389995 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390016 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390035 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390053 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390066 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390154 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390167 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390181 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390217 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390254 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390269 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390284 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390296 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390330 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390343 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390355 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390367 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390381 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390393 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390427 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390439 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390512 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390524 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390538 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390549 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390561 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390574 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390586 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390598 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390633 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390701 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390717 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390729 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390742 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390866 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390880 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390894 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.390954 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391090 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391223 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391240 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391255 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391329 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391342 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391401 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391441 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391454 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391715 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391769 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391783 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391797 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391810 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391854 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391905 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391957 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391976 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.391989 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392041 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392159 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392267 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392317 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392431 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392460 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392503 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392520 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392534 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392548 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392599 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392614 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392643 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392658 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392673 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392705 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.392721 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.393043 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.393091 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.393106 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.393118 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.393136 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394162 4806 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394245 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394263 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394275 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394287 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394325 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394341 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394401 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394415 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394427 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394456 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394522 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394536 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394570 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394584 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394611 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394643 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394656 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394691 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394730 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394745 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394757 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394769 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394810 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394823 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394836 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394854 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394905 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394917 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394953 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394965 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.394993 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395025 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395038 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395049 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395107 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395121 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395140 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395152 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395195 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395207 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395219 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395275 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395327 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395351 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395364 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395373 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395404 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395429 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395440 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395449 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395486 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395512 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395521 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395531 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395555 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395601 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395611 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395619 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395641 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395665 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395675 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395683 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395705 4806 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395714 4806 reconstruct.go:97] "Volume reconstruction finished" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.395720 4806 reconciler.go:26] "Reconciler: start to sync state" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.407441 4806 manager.go:324] Recovery completed Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.418641 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.420167 4806 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.420832 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.420875 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.420893 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.422373 4806 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.422411 4806 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.422434 4806 kubelet.go:2335] "Starting kubelet main sync loop" Dec 04 03:54:45 crc kubenswrapper[4806]: E1204 03:54:45.422523 4806 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.423280 4806 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.423296 4806 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.423313 4806 state_mem.go:36] "Initialized new in-memory state store" Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.423911 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:54:45 crc kubenswrapper[4806]: E1204 03:54:45.424054 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.432582 4806 policy_none.go:49] "None policy: Start" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.434839 4806 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.434870 4806 state_mem.go:35] "Initializing new in-memory state store" Dec 04 03:54:45 crc kubenswrapper[4806]: E1204 03:54:45.452069 4806 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.491467 4806 manager.go:334] "Starting Device Plugin manager" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.491646 4806 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.491961 4806 server.go:79] "Starting device plugin registration server" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.493654 4806 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.493699 4806 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.494084 4806 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.494245 4806 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.494258 4806 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 04 03:54:45 crc kubenswrapper[4806]: E1204 03:54:45.503197 4806 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.523236 4806 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.523345 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.524200 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.524245 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.524255 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.524365 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.524842 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.524884 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.525255 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.525280 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.525289 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.525359 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.525477 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.525526 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.525861 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.525880 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.525888 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.525938 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.526078 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.526087 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.526089 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.526177 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.528313 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.528338 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.528359 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.528594 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.528733 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.528783 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.529617 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.529645 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.529655 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.529759 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.529797 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.529818 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.530120 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.530170 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.530371 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.530774 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.530797 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.530807 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.531761 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.531788 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.531798 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.532357 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.532382 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.532391 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:45 crc kubenswrapper[4806]: E1204 03:54:45.564242 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="400ms" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.593871 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.594816 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.594857 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.594867 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.594892 4806 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 03:54:45 crc kubenswrapper[4806]: E1204 03:54:45.595338 4806 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.598702 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.598730 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.598752 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.598768 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.598788 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.598833 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.598881 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.598909 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.598959 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.598986 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.599007 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.599021 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.599044 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.599061 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.599075 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700035 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700122 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700152 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700179 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700200 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700217 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700201 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700251 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700289 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700300 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700307 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700343 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700348 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700365 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700367 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700390 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700323 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700392 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700412 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700418 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700444 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700140 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700465 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700446 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700485 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700453 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700507 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700524 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700614 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.700226 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.795642 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.796896 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.796941 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.796958 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.796984 4806 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 03:54:45 crc kubenswrapper[4806]: E1204 03:54:45.797311 4806 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.853817 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.858678 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.875523 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.889659 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-3e9b274ef4de89960573c330c21caa3a4fd559fb5eb444f8b9939d99cf5ebd0c WatchSource:0}: Error finding container 3e9b274ef4de89960573c330c21caa3a4fd559fb5eb444f8b9939d99cf5ebd0c: Status 404 returned error can't find the container with id 3e9b274ef4de89960573c330c21caa3a4fd559fb5eb444f8b9939d99cf5ebd0c Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.890002 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-dda194c462790ba888feb58845ee40ca56f786cd9bbc59b02d5edcc3d65555ec WatchSource:0}: Error finding container dda194c462790ba888feb58845ee40ca56f786cd9bbc59b02d5edcc3d65555ec: Status 404 returned error can't find the container with id dda194c462790ba888feb58845ee40ca56f786cd9bbc59b02d5edcc3d65555ec Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.890206 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.895065 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-e4b2d4b6f50652d00e11f2afb3e3b4e92ae355f191db47bd49ed29a252f22a7e WatchSource:0}: Error finding container e4b2d4b6f50652d00e11f2afb3e3b4e92ae355f191db47bd49ed29a252f22a7e: Status 404 returned error can't find the container with id e4b2d4b6f50652d00e11f2afb3e3b4e92ae355f191db47bd49ed29a252f22a7e Dec 04 03:54:45 crc kubenswrapper[4806]: I1204 03:54:45.897014 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.904989 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-c88d5bc36b6c13fae152d03fae8165a0f5351d46bc55b35b2d9ce617a35fd2ad WatchSource:0}: Error finding container c88d5bc36b6c13fae152d03fae8165a0f5351d46bc55b35b2d9ce617a35fd2ad: Status 404 returned error can't find the container with id c88d5bc36b6c13fae152d03fae8165a0f5351d46bc55b35b2d9ce617a35fd2ad Dec 04 03:54:45 crc kubenswrapper[4806]: W1204 03:54:45.914746 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-849ef139951907dd17b10fa0f491b9ca0f7607ff9a890912d8aed7b63251b0b5 WatchSource:0}: Error finding container 849ef139951907dd17b10fa0f491b9ca0f7607ff9a890912d8aed7b63251b0b5: Status 404 returned error can't find the container with id 849ef139951907dd17b10fa0f491b9ca0f7607ff9a890912d8aed7b63251b0b5 Dec 04 03:54:45 crc kubenswrapper[4806]: E1204 03:54:45.965145 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="800ms" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.198217 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.199631 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.199653 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.199663 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.199682 4806 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 03:54:46 crc kubenswrapper[4806]: E1204 03:54:46.200097 4806 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Dec 04 03:54:46 crc kubenswrapper[4806]: W1204 03:54:46.340572 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:54:46 crc kubenswrapper[4806]: E1204 03:54:46.340644 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.346150 4806 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.347187 4806 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 02:08:42.159681454 +0000 UTC Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.347227 4806 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 358h13m55.812457195s for next certificate rotation Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.427630 4806 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648" exitCode=0 Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.427709 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648"} Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.427797 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dda194c462790ba888feb58845ee40ca56f786cd9bbc59b02d5edcc3d65555ec"} Dec 04 03:54:46 crc kubenswrapper[4806]: W1204 03:54:46.427783 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:54:46 crc kubenswrapper[4806]: E1204 03:54:46.427853 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.427894 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.428822 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.428858 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.428869 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.429877 4806 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2" exitCode=0 Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.429961 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2"} Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.430012 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"849ef139951907dd17b10fa0f491b9ca0f7607ff9a890912d8aed7b63251b0b5"} Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.430125 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.431323 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.431344 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.431351 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.432243 4806 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="966a63a12a48455423a3f77e6055886b344e0215cc16c2a2d89bdf1ed967f06d" exitCode=0 Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.432300 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"966a63a12a48455423a3f77e6055886b344e0215cc16c2a2d89bdf1ed967f06d"} Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.432320 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c88d5bc36b6c13fae152d03fae8165a0f5351d46bc55b35b2d9ce617a35fd2ad"} Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.432401 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.433162 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.433186 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.433198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.433341 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.434097 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.434119 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.434132 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.434697 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28"} Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.434729 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e4b2d4b6f50652d00e11f2afb3e3b4e92ae355f191db47bd49ed29a252f22a7e"} Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.437178 4806 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7" exitCode=0 Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.437205 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7"} Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.437224 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3e9b274ef4de89960573c330c21caa3a4fd559fb5eb444f8b9939d99cf5ebd0c"} Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.437278 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.437952 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.437997 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:46 crc kubenswrapper[4806]: I1204 03:54:46.438008 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:46 crc kubenswrapper[4806]: W1204 03:54:46.747727 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:54:46 crc kubenswrapper[4806]: E1204 03:54:46.747807 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Dec 04 03:54:46 crc kubenswrapper[4806]: E1204 03:54:46.768491 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="1.6s" Dec 04 03:54:46 crc kubenswrapper[4806]: W1204 03:54:46.800472 4806 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:54:46 crc kubenswrapper[4806]: E1204 03:54:46.800565 4806 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.000835 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.003367 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.003480 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.003489 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.003508 4806 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 03:54:47 crc kubenswrapper[4806]: E1204 03:54:47.004205 4806 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.442888 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff"} Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.442966 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10"} Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.442986 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e"} Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.442983 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.443614 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.443638 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.443648 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.444762 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"63d42708d1f2974b0a6eb579a33c340969fa6df1100355754cb12ceb117f305f"} Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.444841 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.445643 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.445671 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.445679 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.450117 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9"} Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.450162 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617"} Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.450175 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44"} Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.450185 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da"} Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.452645 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e4e4be31ada4276e640bcf04ab82968c378cfaa920bc82ce54d925d457c51585"} Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.452685 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fd1df0baab4bc2ca707440eaa20c66a5b003a783d48ac8ed0c66d185a6e4d116"} Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.452701 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"59760c3aea1ddf4bdfeda36dd7e6e20bb56985e39f27864475a064d0061325de"} Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.452799 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.453657 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.453685 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.453698 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.456394 4806 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e851dfa78c781cb95e3e5a3ea73c38d0c4f1059eebace9a6e17b78a2fc879b2f" exitCode=0 Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.456431 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e851dfa78c781cb95e3e5a3ea73c38d0c4f1059eebace9a6e17b78a2fc879b2f"} Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.456536 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.457354 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.457381 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:47 crc kubenswrapper[4806]: I1204 03:54:47.457392 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.461462 4806 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="90ad1f66cf96de520eed60dd21a9817ee3525980fcb43c1f7fd942085c3b49aa" exitCode=0 Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.461541 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"90ad1f66cf96de520eed60dd21a9817ee3525980fcb43c1f7fd942085c3b49aa"} Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.461672 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.462355 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.462393 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.462407 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.465122 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.465134 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909"} Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.465122 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.468545 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.468579 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.468588 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.468555 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.468825 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.468905 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.604672 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.606294 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.606345 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.606361 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:48 crc kubenswrapper[4806]: I1204 03:54:48.606396 4806 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.170661 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.170785 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.171684 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.171717 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.171729 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.470289 4806 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.470330 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.470700 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9ec6de5f42f8d12b7fb0770059775b12e82faad5a81e2c0e9ef85f6212cffae7"} Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.470726 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6f2f51aea4a8f01c7d96bc057fef542163504caf5a97320bf382e60eabaec8cb"} Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.470736 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"87e159737687699b84f24ff0e37415ca1728f856572a87b95e5588537606f0ac"} Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.470743 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1cc26e5377fcd914affc21e46397fd6ebe234488cc0eadac1c4e9b543301fca2"} Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.470751 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"25f41c8816799ba262455a943cacc42000e10bcea3a3f90ebca0036e42d5927b"} Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.470813 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.471386 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.471410 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.471426 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.472067 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.472083 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:49 crc kubenswrapper[4806]: I1204 03:54:49.472091 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.146091 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.146312 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.147752 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.147800 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.147816 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.154533 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.402530 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.474328 4806 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.474396 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.474399 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.475856 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.475916 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.475990 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.476504 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.476678 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.476823 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:50 crc kubenswrapper[4806]: I1204 03:54:50.899284 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:54:51 crc kubenswrapper[4806]: I1204 03:54:51.269154 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 04 03:54:51 crc kubenswrapper[4806]: I1204 03:54:51.269487 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:51 crc kubenswrapper[4806]: I1204 03:54:51.273751 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:51 crc kubenswrapper[4806]: I1204 03:54:51.274007 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:51 crc kubenswrapper[4806]: I1204 03:54:51.274105 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:51 crc kubenswrapper[4806]: I1204 03:54:51.476493 4806 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 03:54:51 crc kubenswrapper[4806]: I1204 03:54:51.477197 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:51 crc kubenswrapper[4806]: I1204 03:54:51.478574 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:51 crc kubenswrapper[4806]: I1204 03:54:51.478619 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:51 crc kubenswrapper[4806]: I1204 03:54:51.478633 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:51 crc kubenswrapper[4806]: I1204 03:54:51.610089 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:54:51 crc kubenswrapper[4806]: I1204 03:54:51.610334 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:51 crc kubenswrapper[4806]: I1204 03:54:51.612080 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:51 crc kubenswrapper[4806]: I1204 03:54:51.612133 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:51 crc kubenswrapper[4806]: I1204 03:54:51.612143 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:53 crc kubenswrapper[4806]: I1204 03:54:53.051911 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 04 03:54:53 crc kubenswrapper[4806]: I1204 03:54:53.052150 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:53 crc kubenswrapper[4806]: I1204 03:54:53.053904 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:53 crc kubenswrapper[4806]: I1204 03:54:53.053986 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:53 crc kubenswrapper[4806]: I1204 03:54:53.054006 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:53 crc kubenswrapper[4806]: I1204 03:54:53.351768 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:54:53 crc kubenswrapper[4806]: I1204 03:54:53.352106 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:53 crc kubenswrapper[4806]: I1204 03:54:53.354341 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:53 crc kubenswrapper[4806]: I1204 03:54:53.354410 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:53 crc kubenswrapper[4806]: I1204 03:54:53.354428 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:53 crc kubenswrapper[4806]: I1204 03:54:53.390681 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:54:53 crc kubenswrapper[4806]: I1204 03:54:53.390969 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:53 crc kubenswrapper[4806]: I1204 03:54:53.392461 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:53 crc kubenswrapper[4806]: I1204 03:54:53.392548 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:53 crc kubenswrapper[4806]: I1204 03:54:53.392599 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:55 crc kubenswrapper[4806]: E1204 03:54:55.503675 4806 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 04 03:54:56 crc kubenswrapper[4806]: I1204 03:54:56.069638 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:54:56 crc kubenswrapper[4806]: I1204 03:54:56.069919 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:56 crc kubenswrapper[4806]: I1204 03:54:56.071505 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:56 crc kubenswrapper[4806]: I1204 03:54:56.071563 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:56 crc kubenswrapper[4806]: I1204 03:54:56.071577 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:56 crc kubenswrapper[4806]: I1204 03:54:56.075321 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:54:56 crc kubenswrapper[4806]: I1204 03:54:56.487640 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:54:56 crc kubenswrapper[4806]: I1204 03:54:56.488497 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:54:56 crc kubenswrapper[4806]: I1204 03:54:56.488553 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:54:56 crc kubenswrapper[4806]: I1204 03:54:56.488570 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:54:57 crc kubenswrapper[4806]: I1204 03:54:57.346641 4806 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 04 03:54:58 crc kubenswrapper[4806]: I1204 03:54:58.128338 4806 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 04 03:54:58 crc kubenswrapper[4806]: I1204 03:54:58.128416 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 04 03:54:58 crc kubenswrapper[4806]: I1204 03:54:58.401816 4806 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]log ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]etcd ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/generic-apiserver-start-informers ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/priority-and-fairness-filter ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/start-apiextensions-informers ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/start-apiextensions-controllers ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/crd-informer-synced ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/start-system-namespaces-controller ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 04 03:54:58 crc kubenswrapper[4806]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 04 03:54:58 crc kubenswrapper[4806]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/bootstrap-controller ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/start-kube-aggregator-informers ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/apiservice-registration-controller ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/apiservice-discovery-controller ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]autoregister-completion ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/apiservice-openapi-controller ok Dec 04 03:54:58 crc kubenswrapper[4806]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 04 03:54:58 crc kubenswrapper[4806]: livez check failed Dec 04 03:54:58 crc kubenswrapper[4806]: I1204 03:54:58.401870 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:54:59 crc kubenswrapper[4806]: I1204 03:54:59.070592 4806 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 03:54:59 crc kubenswrapper[4806]: I1204 03:54:59.070695 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 03:55:01 crc kubenswrapper[4806]: I1204 03:55:01.295707 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 04 03:55:01 crc kubenswrapper[4806]: I1204 03:55:01.295963 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:55:01 crc kubenswrapper[4806]: I1204 03:55:01.297416 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:01 crc kubenswrapper[4806]: I1204 03:55:01.297466 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:01 crc kubenswrapper[4806]: I1204 03:55:01.297480 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:01 crc kubenswrapper[4806]: I1204 03:55:01.311918 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 04 03:55:01 crc kubenswrapper[4806]: I1204 03:55:01.500511 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:55:01 crc kubenswrapper[4806]: I1204 03:55:01.501601 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:01 crc kubenswrapper[4806]: I1204 03:55:01.501642 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:01 crc kubenswrapper[4806]: I1204 03:55:01.501654 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:01 crc kubenswrapper[4806]: I1204 03:55:01.611415 4806 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 04 03:55:01 crc kubenswrapper[4806]: I1204 03:55:01.611495 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 04 03:55:02 crc kubenswrapper[4806]: I1204 03:55:02.840980 4806 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 04 03:55:02 crc kubenswrapper[4806]: I1204 03:55:02.841032 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.129179 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.149741 4806 trace.go:236] Trace[1358434581]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 03:54:48.296) (total time: 14852ms): Dec 04 03:55:03 crc kubenswrapper[4806]: Trace[1358434581]: ---"Objects listed" error: 14852ms (03:55:03.149) Dec 04 03:55:03 crc kubenswrapper[4806]: Trace[1358434581]: [14.852921925s] [14.852921925s] END Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.149781 4806 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.150209 4806 trace.go:236] Trace[238556337]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 03:54:48.374) (total time: 14775ms): Dec 04 03:55:03 crc kubenswrapper[4806]: Trace[238556337]: ---"Objects listed" error: 14775ms (03:55:03.150) Dec 04 03:55:03 crc kubenswrapper[4806]: Trace[238556337]: [14.775502307s] [14.775502307s] END Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.150244 4806 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.150579 4806 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.150606 4806 trace.go:236] Trace[400284558]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 03:54:49.634) (total time: 13516ms): Dec 04 03:55:03 crc kubenswrapper[4806]: Trace[400284558]: ---"Objects listed" error: 13516ms (03:55:03.150) Dec 04 03:55:03 crc kubenswrapper[4806]: Trace[400284558]: [13.516270958s] [13.516270958s] END Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.150617 4806 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.150878 4806 trace.go:236] Trace[1905531430]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 03:54:49.150) (total time: 14000ms): Dec 04 03:55:03 crc kubenswrapper[4806]: Trace[1905531430]: ---"Objects listed" error: 14000ms (03:55:03.150) Dec 04 03:55:03 crc kubenswrapper[4806]: Trace[1905531430]: [14.000132093s] [14.000132093s] END Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.150901 4806 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.154891 4806 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.347859 4806 apiserver.go:52] "Watching apiserver" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.350087 4806 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.350355 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.350628 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.351119 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.351170 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.351405 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.351448 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.351591 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.352003 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.352034 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.352053 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.352074 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.352095 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.352100 4806 configmap.go:193] Couldn't get configMap openshift-network-operator/iptables-alerter-script: object "openshift-network-operator"/"iptables-alerter-script" not registered Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.352118 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.352138 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.352159 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script podName:d75a4c96-2883-4a0b-bab2-0fab2b6c0b49 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:03.85213485 +0000 UTC m=+18.710647778 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "iptables-alerter-script" (UniqueName: "kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script") pod "iptables-alerter-4ln5h" (UID: "d75a4c96-2883-4a0b-bab2-0fab2b6c0b49") : object "openshift-network-operator"/"iptables-alerter-script" not registered Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.352182 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.352211 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.352232 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.352253 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.352274 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.352660 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.352891 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.352944 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.360124 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.360578 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.361839 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.362976 4806 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.364041 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.365453 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.365617 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.365776 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.365861 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.365869 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.371908 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.373461 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.373566 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.377863 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.395370 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.395643 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.395752 4806 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.395901 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:03.895879378 +0000 UTC m=+18.754392306 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.395560 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.396111 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.396191 4806 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.396315 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:03.896286541 +0000 UTC m=+18.754799469 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.399627 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.403727 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.407473 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.410047 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.413352 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.419441 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.441086 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.442972 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.452499 4806 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.452945 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453000 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453028 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453050 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453075 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453100 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453122 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453147 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453171 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453193 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453213 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453235 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453256 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453277 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453299 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453313 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453323 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453388 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453418 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453439 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453462 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453483 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453506 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453529 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453552 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453574 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453597 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453602 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453620 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453647 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453670 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453694 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453723 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453746 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453768 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453794 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453817 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453838 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453862 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453884 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453906 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453945 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453968 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.453994 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454015 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454031 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454039 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454091 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454099 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454126 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454148 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454171 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454194 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454217 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454238 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454264 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454287 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454308 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454330 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454350 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454372 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454394 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454415 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454436 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454457 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454477 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454497 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454517 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454521 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454540 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454560 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454582 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454602 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454621 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454643 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454666 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454692 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454685 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454735 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454759 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454782 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454803 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454826 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454847 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454868 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454887 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454891 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454950 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454975 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.454998 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455024 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455045 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455067 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455077 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455089 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455110 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455155 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455177 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455200 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455223 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455244 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455267 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455322 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455345 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455366 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455388 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455414 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455436 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455457 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455480 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455503 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455525 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455550 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455562 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455571 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455593 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455618 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455643 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455665 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455710 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455733 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455755 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455777 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455782 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455800 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455824 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455848 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455872 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455895 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.455972 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456148 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456426 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456464 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456491 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456514 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456539 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456565 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456588 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456597 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456614 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456638 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456662 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456685 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456708 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456731 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456754 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456778 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456786 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456802 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456826 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456850 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456874 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456898 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456941 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456964 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.456986 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457012 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457036 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457058 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457082 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457108 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457138 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457165 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457187 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457297 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457324 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457350 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457381 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457408 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457435 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457461 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457482 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457505 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457528 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457553 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457578 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457603 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457638 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457664 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457687 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457710 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457733 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457773 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457797 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457822 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457848 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457871 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457895 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457946 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457975 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458001 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458031 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458058 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458084 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458108 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458133 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458157 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458181 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458207 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458232 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458257 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458285 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458309 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458335 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458360 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458386 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458411 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458436 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458460 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458547 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458604 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458649 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458678 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458842 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458859 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458873 4806 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458886 4806 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458901 4806 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458914 4806 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458945 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458959 4806 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458972 4806 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458985 4806 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458996 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.459009 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.459024 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.459037 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.470354 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.457732 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458267 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.458894 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.459179 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.459256 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.459528 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.459623 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.459693 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.459756 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.460097 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.460140 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.460322 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.460478 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.460437 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.460476 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.460648 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.460722 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.460730 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.460830 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.461018 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.461058 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.461031 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.461257 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.461332 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.461359 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.461460 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.461557 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.461658 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.461860 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.461941 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.462017 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.462095 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.462223 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.462250 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.462508 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.462522 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.462788 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.463007 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.463074 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.463266 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.463449 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.463668 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.463698 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.464058 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.464468 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.464635 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.464782 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.464945 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.465020 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.465095 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.465283 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.465349 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.465502 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.465824 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.465829 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.466104 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.466669 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.466868 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.467170 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.467216 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.467587 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.467643 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.467777 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.467995 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.468002 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.468095 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.468144 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.468264 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.468351 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.468557 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.468562 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.468715 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.468766 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.468888 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.468992 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.469170 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.469205 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.469352 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.469511 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.469752 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.469970 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.470143 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.470718 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.470848 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.471030 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.471089 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.471235 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.471275 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.471514 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.471594 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.471679 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.471780 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.472099 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.472284 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.472430 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.472579 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.473274 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.473428 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.473858 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.474130 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.474326 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.474796 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.475738 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.475894 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.476090 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.476264 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.476418 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.476614 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.476710 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:55:03.976692265 +0000 UTC m=+18.835205203 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.492402 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.489842 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.477083 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.478095 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.479379 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.479512 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.479680 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.479884 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.480069 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.480170 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.480287 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.480338 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.480629 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.480760 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.480865 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.480949 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.481124 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.481239 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.481783 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.482047 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.482196 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.482276 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.482289 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.482474 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.483648 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.483667 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.483694 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.483960 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.483967 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.484100 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.484457 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.484465 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.484886 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.485243 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.485334 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.485535 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.485967 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.486368 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.486810 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.487248 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.487600 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.487696 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.487909 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.487984 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.487998 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.488071 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.488225 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.488414 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.488735 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.488758 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.488958 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.489171 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.489221 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.487865 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.489337 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.489551 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.489803 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.489866 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.489906 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.490670 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.490685 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.490733 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.492948 4806 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.493262 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:03.993241403 +0000 UTC m=+18.851754341 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.493269 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.493340 4806 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.493398 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:03.993379657 +0000 UTC m=+18.851892665 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.494302 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.494873 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.476874 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.496948 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.497575 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.499031 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.503122 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.503660 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.507347 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.509806 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.512620 4806 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.512669 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.516084 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.520683 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.523445 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.533694 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.544601 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.558534 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.559835 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.559884 4806 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.559893 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.559902 4806 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.559911 4806 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.559935 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.559947 4806 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.559960 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.559969 4806 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.559977 4806 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.559985 4806 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.559995 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560003 4806 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560011 4806 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560018 4806 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560026 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560034 4806 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560041 4806 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560051 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560069 4806 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560081 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560097 4806 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560113 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560123 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560134 4806 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560146 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560156 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560165 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560176 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560187 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560196 4806 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560208 4806 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560219 4806 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560230 4806 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560241 4806 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560257 4806 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560270 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560279 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560289 4806 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560300 4806 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560309 4806 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560319 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560329 4806 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560343 4806 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560352 4806 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560361 4806 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560371 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560380 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560390 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560400 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560410 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560420 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560430 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560450 4806 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560462 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560471 4806 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560482 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560494 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560504 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560514 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560522 4806 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560529 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560537 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560544 4806 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560552 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560559 4806 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560567 4806 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560575 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560582 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560590 4806 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560597 4806 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560604 4806 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560611 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560626 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560634 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560641 4806 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560650 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560659 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560667 4806 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560675 4806 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560682 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560690 4806 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560703 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560711 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560719 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560727 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560734 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560742 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560751 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560759 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560766 4806 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560774 4806 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560781 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560789 4806 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560797 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560806 4806 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560814 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560822 4806 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560829 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560837 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560844 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560852 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560859 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560867 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560874 4806 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560882 4806 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560890 4806 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560897 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560905 4806 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560914 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560949 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560969 4806 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560982 4806 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.560993 4806 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561004 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561013 4806 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561023 4806 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561031 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561039 4806 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561049 4806 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561064 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561080 4806 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561092 4806 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561102 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561113 4806 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561124 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561134 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561145 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561155 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561164 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561172 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561180 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561188 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561196 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561204 4806 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561211 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561229 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561245 4806 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561258 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561268 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561278 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561289 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561299 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561312 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561322 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561333 4806 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561343 4806 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561353 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561362 4806 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561372 4806 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561409 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561421 4806 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561432 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561441 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561496 4806 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561507 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561517 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561527 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561537 4806 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561548 4806 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561559 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561571 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561583 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561594 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561605 4806 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561618 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561632 4806 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561643 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561654 4806 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561664 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561676 4806 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561694 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561706 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561716 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561732 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561743 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561754 4806 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561765 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561776 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.561788 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.573411 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.583997 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.593601 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.606015 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.615142 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.662951 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.673878 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.727144 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.739901 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.757274 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.778791 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.794584 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.837138 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.865506 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.865952 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.869584 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.893634 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.966740 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.967087 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.966905 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.967312 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.967380 4806 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.967474 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:04.967461287 +0000 UTC m=+19.825974215 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.967192 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.967594 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.967651 4806 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:03 crc kubenswrapper[4806]: E1204 03:55:03.967727 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:04.967719885 +0000 UTC m=+19.826232813 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:03 crc kubenswrapper[4806]: I1204 03:55:03.990669 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 03:55:04 crc kubenswrapper[4806]: W1204 03:55:04.001139 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-86520cc9a4a25bb9c418b189ac4bdf6272530335c0ddcbd48d963a9c468714cb WatchSource:0}: Error finding container 86520cc9a4a25bb9c418b189ac4bdf6272530335c0ddcbd48d963a9c468714cb: Status 404 returned error can't find the container with id 86520cc9a4a25bb9c418b189ac4bdf6272530335c0ddcbd48d963a9c468714cb Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.068106 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.068180 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.068210 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:04 crc kubenswrapper[4806]: E1204 03:55:04.068261 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:55:05.06824048 +0000 UTC m=+19.926753418 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:55:04 crc kubenswrapper[4806]: E1204 03:55:04.068303 4806 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 03:55:04 crc kubenswrapper[4806]: E1204 03:55:04.068345 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:05.068334893 +0000 UTC m=+19.926847821 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 03:55:04 crc kubenswrapper[4806]: E1204 03:55:04.068347 4806 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 03:55:04 crc kubenswrapper[4806]: E1204 03:55:04.068380 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:05.068371674 +0000 UTC m=+19.926884602 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.508652 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848"} Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.509049 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2"} Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.509063 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"704111df658fc18f4eb78827ab0c2a9e467d39c2ffec42cb0efff209d2896dfd"} Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.510207 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4"} Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.510322 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8d36439077fdf53a9d39ecfb2d2ce64bf3d3faeeac532c28bef4c148d3d99274"} Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.511824 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"86520cc9a4a25bb9c418b189ac4bdf6272530335c0ddcbd48d963a9c468714cb"} Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.550976 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.564206 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.579835 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.593176 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.612686 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.629548 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.647130 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.660879 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.678385 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.691793 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.708163 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.721167 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.733576 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.746705 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.938266 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-qdpsj"] Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.938619 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-qdpsj" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.940618 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.947229 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.947624 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.947748 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.963185 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:04 crc kubenswrapper[4806]: E1204 03:55:04.981377 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 03:55:04 crc kubenswrapper[4806]: E1204 03:55:04.981409 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 03:55:04 crc kubenswrapper[4806]: E1204 03:55:04.981421 4806 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:04 crc kubenswrapper[4806]: E1204 03:55:04.981472 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:06.981457176 +0000 UTC m=+21.839970094 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.981497 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.981553 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v229f\" (UniqueName: \"kubernetes.io/projected/5a576e68-334a-4a4f-aae0-888b6293dc41-kube-api-access-v229f\") pod \"node-ca-qdpsj\" (UID: \"5a576e68-334a-4a4f-aae0-888b6293dc41\") " pod="openshift-image-registry/node-ca-qdpsj" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.981574 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.981603 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5a576e68-334a-4a4f-aae0-888b6293dc41-host\") pod \"node-ca-qdpsj\" (UID: \"5a576e68-334a-4a4f-aae0-888b6293dc41\") " pod="openshift-image-registry/node-ca-qdpsj" Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.981619 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5a576e68-334a-4a4f-aae0-888b6293dc41-serviceca\") pod \"node-ca-qdpsj\" (UID: \"5a576e68-334a-4a4f-aae0-888b6293dc41\") " pod="openshift-image-registry/node-ca-qdpsj" Dec 04 03:55:04 crc kubenswrapper[4806]: E1204 03:55:04.981727 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 03:55:04 crc kubenswrapper[4806]: E1204 03:55:04.981737 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 03:55:04 crc kubenswrapper[4806]: E1204 03:55:04.981749 4806 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:04 crc kubenswrapper[4806]: E1204 03:55:04.981775 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:06.981764465 +0000 UTC m=+21.840277393 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:04 crc kubenswrapper[4806]: I1204 03:55:04.996048 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.045525 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.082453 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.082530 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.082559 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:05 crc kubenswrapper[4806]: E1204 03:55:05.082591 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:55:07.082571339 +0000 UTC m=+21.941084267 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.082631 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v229f\" (UniqueName: \"kubernetes.io/projected/5a576e68-334a-4a4f-aae0-888b6293dc41-kube-api-access-v229f\") pod \"node-ca-qdpsj\" (UID: \"5a576e68-334a-4a4f-aae0-888b6293dc41\") " pod="openshift-image-registry/node-ca-qdpsj" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.082655 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5a576e68-334a-4a4f-aae0-888b6293dc41-host\") pod \"node-ca-qdpsj\" (UID: \"5a576e68-334a-4a4f-aae0-888b6293dc41\") " pod="openshift-image-registry/node-ca-qdpsj" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.082675 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5a576e68-334a-4a4f-aae0-888b6293dc41-serviceca\") pod \"node-ca-qdpsj\" (UID: \"5a576e68-334a-4a4f-aae0-888b6293dc41\") " pod="openshift-image-registry/node-ca-qdpsj" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.082721 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5a576e68-334a-4a4f-aae0-888b6293dc41-host\") pod \"node-ca-qdpsj\" (UID: \"5a576e68-334a-4a4f-aae0-888b6293dc41\") " pod="openshift-image-registry/node-ca-qdpsj" Dec 04 03:55:05 crc kubenswrapper[4806]: E1204 03:55:05.082660 4806 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 03:55:05 crc kubenswrapper[4806]: E1204 03:55:05.083058 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:07.083046483 +0000 UTC m=+21.941559421 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 03:55:05 crc kubenswrapper[4806]: E1204 03:55:05.082691 4806 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 03:55:05 crc kubenswrapper[4806]: E1204 03:55:05.083101 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:07.083091494 +0000 UTC m=+21.941604422 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.083710 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5a576e68-334a-4a4f-aae0-888b6293dc41-serviceca\") pod \"node-ca-qdpsj\" (UID: \"5a576e68-334a-4a4f-aae0-888b6293dc41\") " pod="openshift-image-registry/node-ca-qdpsj" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.124402 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.141203 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v229f\" (UniqueName: \"kubernetes.io/projected/5a576e68-334a-4a4f-aae0-888b6293dc41-kube-api-access-v229f\") pod \"node-ca-qdpsj\" (UID: \"5a576e68-334a-4a4f-aae0-888b6293dc41\") " pod="openshift-image-registry/node-ca-qdpsj" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.204014 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.241003 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.250692 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-qdpsj" Dec 04 03:55:05 crc kubenswrapper[4806]: W1204 03:55:05.263251 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a576e68_334a_4a4f_aae0_888b6293dc41.slice/crio-b2f5d98871f6fd32dbe832cf8f3ca8d164318f39b8ec43fda21a9eee253001e3 WatchSource:0}: Error finding container b2f5d98871f6fd32dbe832cf8f3ca8d164318f39b8ec43fda21a9eee253001e3: Status 404 returned error can't find the container with id b2f5d98871f6fd32dbe832cf8f3ca8d164318f39b8ec43fda21a9eee253001e3 Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.278277 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.309787 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.412846 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-hqps5"] Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.413152 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hqps5" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.416320 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.416504 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.416580 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.423113 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.423170 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:05 crc kubenswrapper[4806]: E1204 03:55:05.423242 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.423113 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:05 crc kubenswrapper[4806]: E1204 03:55:05.423317 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:05 crc kubenswrapper[4806]: E1204 03:55:05.423492 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.427342 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.428193 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.429480 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.430146 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.431139 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.431601 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.432230 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.433181 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.433799 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.434841 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.435387 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.439065 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.439738 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.440248 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.441239 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.441736 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.442738 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.443132 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.443667 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.444576 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.447805 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.448326 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.449242 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.449641 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.450695 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.451194 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.451752 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.452941 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.453402 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.454345 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.454780 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.455624 4806 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.455719 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.459416 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.460266 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.461377 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.462972 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.463592 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.465170 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.466150 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.466857 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.468219 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.468703 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.469647 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.470556 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.471585 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.472067 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.474291 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.474909 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.476012 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.476453 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.477327 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.478611 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.479591 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.480339 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.481508 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.485882 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8a1fde59-696b-48ad-8fce-491ab2ee8454-hosts-file\") pod \"node-resolver-hqps5\" (UID: \"8a1fde59-696b-48ad-8fce-491ab2ee8454\") " pod="openshift-dns/node-resolver-hqps5" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.485942 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94bvt\" (UniqueName: \"kubernetes.io/projected/8a1fde59-696b-48ad-8fce-491ab2ee8454-kube-api-access-94bvt\") pod \"node-resolver-hqps5\" (UID: \"8a1fde59-696b-48ad-8fce-491ab2ee8454\") " pod="openshift-dns/node-resolver-hqps5" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.494898 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.520048 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-qdpsj" event={"ID":"5a576e68-334a-4a4f-aae0-888b6293dc41","Type":"ContainerStarted","Data":"b2f5d98871f6fd32dbe832cf8f3ca8d164318f39b8ec43fda21a9eee253001e3"} Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.530668 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.559758 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.574656 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.587245 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8a1fde59-696b-48ad-8fce-491ab2ee8454-hosts-file\") pod \"node-resolver-hqps5\" (UID: \"8a1fde59-696b-48ad-8fce-491ab2ee8454\") " pod="openshift-dns/node-resolver-hqps5" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.587532 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94bvt\" (UniqueName: \"kubernetes.io/projected/8a1fde59-696b-48ad-8fce-491ab2ee8454-kube-api-access-94bvt\") pod \"node-resolver-hqps5\" (UID: \"8a1fde59-696b-48ad-8fce-491ab2ee8454\") " pod="openshift-dns/node-resolver-hqps5" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.587389 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8a1fde59-696b-48ad-8fce-491ab2ee8454-hosts-file\") pod \"node-resolver-hqps5\" (UID: \"8a1fde59-696b-48ad-8fce-491ab2ee8454\") " pod="openshift-dns/node-resolver-hqps5" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.616823 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.617412 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94bvt\" (UniqueName: \"kubernetes.io/projected/8a1fde59-696b-48ad-8fce-491ab2ee8454-kube-api-access-94bvt\") pod \"node-resolver-hqps5\" (UID: \"8a1fde59-696b-48ad-8fce-491ab2ee8454\") " pod="openshift-dns/node-resolver-hqps5" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.637495 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.647461 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.658534 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.676177 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.699679 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.713787 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.726029 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hqps5" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.726520 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.740763 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: W1204 03:55:05.744637 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a1fde59_696b_48ad_8fce_491ab2ee8454.slice/crio-ed835111d73bc556e5fd76bcd8558358b35988df718919c772838ee6fb9870ca WatchSource:0}: Error finding container ed835111d73bc556e5fd76bcd8558358b35988df718919c772838ee6fb9870ca: Status 404 returned error can't find the container with id ed835111d73bc556e5fd76bcd8558358b35988df718919c772838ee6fb9870ca Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.754399 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.778341 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.832079 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.862674 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-zpps8"] Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.863306 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.867681 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.867704 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.867858 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-mfj6c"] Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.868335 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.869601 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-dsclk"] Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.870086 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.878754 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.879254 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.879391 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.880619 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.880761 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.880857 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.880881 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.881149 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.881233 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.889770 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-system-cni-dir\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.889797 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-var-lib-cni-bin\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.889824 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-hostroot\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.889842 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d0f9623b-44ce-43c2-b990-a75db2d17ff8-mcd-auth-proxy-config\") pod \"machine-config-daemon-zpps8\" (UID: \"d0f9623b-44ce-43c2-b990-a75db2d17ff8\") " pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.889862 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/10aef7a5-56dc-4742-87d3-9c87074eb77a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.889877 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j54q\" (UniqueName: \"kubernetes.io/projected/10aef7a5-56dc-4742-87d3-9c87074eb77a-kube-api-access-2j54q\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.889891 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-cnibin\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.889913 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-var-lib-kubelet\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.889954 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z858j\" (UniqueName: \"kubernetes.io/projected/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-kube-api-access-z858j\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.889975 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/10aef7a5-56dc-4742-87d3-9c87074eb77a-system-cni-dir\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.889991 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-multus-socket-dir-parent\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890011 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-multus-cni-dir\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890023 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-multus-conf-dir\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890037 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-etc-kubernetes\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890071 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmqn8\" (UniqueName: \"kubernetes.io/projected/d0f9623b-44ce-43c2-b990-a75db2d17ff8-kube-api-access-fmqn8\") pod \"machine-config-daemon-zpps8\" (UID: \"d0f9623b-44ce-43c2-b990-a75db2d17ff8\") " pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890089 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d0f9623b-44ce-43c2-b990-a75db2d17ff8-proxy-tls\") pod \"machine-config-daemon-zpps8\" (UID: \"d0f9623b-44ce-43c2-b990-a75db2d17ff8\") " pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890107 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/10aef7a5-56dc-4742-87d3-9c87074eb77a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890120 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/10aef7a5-56dc-4742-87d3-9c87074eb77a-os-release\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890135 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-var-lib-cni-multus\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890148 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-os-release\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890161 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-run-multus-certs\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890174 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d0f9623b-44ce-43c2-b990-a75db2d17ff8-rootfs\") pod \"machine-config-daemon-zpps8\" (UID: \"d0f9623b-44ce-43c2-b990-a75db2d17ff8\") " pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890188 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-run-k8s-cni-cncf-io\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890203 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-run-netns\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890219 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-cni-binary-copy\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890236 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/10aef7a5-56dc-4742-87d3-9c87074eb77a-cnibin\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890252 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/10aef7a5-56dc-4742-87d3-9c87074eb77a-cni-binary-copy\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.890273 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-multus-daemon-config\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.892420 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.896153 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.922458 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.959219 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.986243 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991321 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-multus-socket-dir-parent\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991363 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmqn8\" (UniqueName: \"kubernetes.io/projected/d0f9623b-44ce-43c2-b990-a75db2d17ff8-kube-api-access-fmqn8\") pod \"machine-config-daemon-zpps8\" (UID: \"d0f9623b-44ce-43c2-b990-a75db2d17ff8\") " pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991381 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-multus-cni-dir\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991396 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-multus-conf-dir\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991415 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-etc-kubernetes\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991433 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/10aef7a5-56dc-4742-87d3-9c87074eb77a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991449 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d0f9623b-44ce-43c2-b990-a75db2d17ff8-proxy-tls\") pod \"machine-config-daemon-zpps8\" (UID: \"d0f9623b-44ce-43c2-b990-a75db2d17ff8\") " pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991464 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/10aef7a5-56dc-4742-87d3-9c87074eb77a-os-release\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991481 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-var-lib-cni-multus\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991497 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-os-release\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991494 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-multus-socket-dir-parent\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991550 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-run-multus-certs\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991513 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-run-multus-certs\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991587 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d0f9623b-44ce-43c2-b990-a75db2d17ff8-rootfs\") pod \"machine-config-daemon-zpps8\" (UID: \"d0f9623b-44ce-43c2-b990-a75db2d17ff8\") " pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991603 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-run-k8s-cni-cncf-io\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991610 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-multus-cni-dir\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991640 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-run-netns\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991619 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-run-netns\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992006 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-cni-binary-copy\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992032 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/10aef7a5-56dc-4742-87d3-9c87074eb77a-cnibin\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992041 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/10aef7a5-56dc-4742-87d3-9c87074eb77a-os-release\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992049 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/10aef7a5-56dc-4742-87d3-9c87074eb77a-cni-binary-copy\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991837 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/10aef7a5-56dc-4742-87d3-9c87074eb77a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991870 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-run-k8s-cni-cncf-io\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991966 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-os-release\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991668 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-var-lib-cni-multus\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991720 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d0f9623b-44ce-43c2-b990-a75db2d17ff8-rootfs\") pod \"machine-config-daemon-zpps8\" (UID: \"d0f9623b-44ce-43c2-b990-a75db2d17ff8\") " pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991679 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-multus-conf-dir\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.991701 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-etc-kubernetes\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992092 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/10aef7a5-56dc-4742-87d3-9c87074eb77a-cnibin\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992114 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-multus-daemon-config\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992218 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-system-cni-dir\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992236 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-var-lib-cni-bin\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992273 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-hostroot\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992304 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d0f9623b-44ce-43c2-b990-a75db2d17ff8-mcd-auth-proxy-config\") pod \"machine-config-daemon-zpps8\" (UID: \"d0f9623b-44ce-43c2-b990-a75db2d17ff8\") " pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992319 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j54q\" (UniqueName: \"kubernetes.io/projected/10aef7a5-56dc-4742-87d3-9c87074eb77a-kube-api-access-2j54q\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992334 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-cnibin\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992363 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/10aef7a5-56dc-4742-87d3-9c87074eb77a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992383 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/10aef7a5-56dc-4742-87d3-9c87074eb77a-system-cni-dir\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992396 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-var-lib-kubelet\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992411 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z858j\" (UniqueName: \"kubernetes.io/projected/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-kube-api-access-z858j\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992474 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-cnibin\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992502 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-system-cni-dir\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992697 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-cni-binary-copy\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992727 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/10aef7a5-56dc-4742-87d3-9c87074eb77a-system-cni-dir\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992798 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/10aef7a5-56dc-4742-87d3-9c87074eb77a-cni-binary-copy\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.992813 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-multus-daemon-config\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.993099 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d0f9623b-44ce-43c2-b990-a75db2d17ff8-mcd-auth-proxy-config\") pod \"machine-config-daemon-zpps8\" (UID: \"d0f9623b-44ce-43c2-b990-a75db2d17ff8\") " pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.993559 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/10aef7a5-56dc-4742-87d3-9c87074eb77a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.993587 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-var-lib-kubelet\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.993601 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-hostroot\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.993613 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-host-var-lib-cni-bin\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:05 crc kubenswrapper[4806]: I1204 03:55:05.995104 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d0f9623b-44ce-43c2-b990-a75db2d17ff8-proxy-tls\") pod \"machine-config-daemon-zpps8\" (UID: \"d0f9623b-44ce-43c2-b990-a75db2d17ff8\") " pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.027173 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j54q\" (UniqueName: \"kubernetes.io/projected/10aef7a5-56dc-4742-87d3-9c87074eb77a-kube-api-access-2j54q\") pod \"multus-additional-cni-plugins-mfj6c\" (UID: \"10aef7a5-56dc-4742-87d3-9c87074eb77a\") " pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.030468 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmqn8\" (UniqueName: \"kubernetes.io/projected/d0f9623b-44ce-43c2-b990-a75db2d17ff8-kube-api-access-fmqn8\") pod \"machine-config-daemon-zpps8\" (UID: \"d0f9623b-44ce-43c2-b990-a75db2d17ff8\") " pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.034197 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.034619 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z858j\" (UniqueName: \"kubernetes.io/projected/4f13bcfc-c287-40fd-b8f6-e4bbea8c6577-kube-api-access-z858j\") pod \"multus-dsclk\" (UID: \"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\") " pod="openshift-multus/multus-dsclk" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.059470 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.075317 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.078285 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.084526 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.092609 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.105059 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.119483 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.134605 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.149691 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.162165 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.173172 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.183723 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.192774 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.193956 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.202118 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" Dec 04 03:55:06 crc kubenswrapper[4806]: W1204 03:55:06.205201 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0f9623b_44ce_43c2_b990_a75db2d17ff8.slice/crio-bbb9b8a48c0d140661a074a51a35137fffe28cf94a51d78e733f853960f411e1 WatchSource:0}: Error finding container bbb9b8a48c0d140661a074a51a35137fffe28cf94a51d78e733f853960f411e1: Status 404 returned error can't find the container with id bbb9b8a48c0d140661a074a51a35137fffe28cf94a51d78e733f853960f411e1 Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.212190 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.214262 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-dsclk" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.234455 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: W1204 03:55:06.237651 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f13bcfc_c287_40fd_b8f6_e4bbea8c6577.slice/crio-85045d70fbe69286388ffd26bcde03943cd51cad7ea143e116359b2cf58c5acb WatchSource:0}: Error finding container 85045d70fbe69286388ffd26bcde03943cd51cad7ea143e116359b2cf58c5acb: Status 404 returned error can't find the container with id 85045d70fbe69286388ffd26bcde03943cd51cad7ea143e116359b2cf58c5acb Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.251774 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.269614 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.282017 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.293451 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.301666 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ljxsg"] Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.308915 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.313048 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.314258 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.315039 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.315389 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.315695 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.317347 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.317896 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.318021 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.329286 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.343700 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.355616 4806 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.359289 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.359512 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.359541 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.360012 4806 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.369295 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.374901 4806 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.375134 4806 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.375945 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.375968 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.375977 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.375992 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.376004 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:06Z","lastTransitionTime":"2025-12-04T03:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396491 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-systemd\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396538 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-cni-netd\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396571 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396593 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-kubelet\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396613 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-systemd-units\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396634 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-etc-openvswitch\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396654 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-node-log\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396674 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-log-socket\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396691 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovnkube-config\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396710 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-run-ovn-kubernetes\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396680 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396733 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-var-lib-openvswitch\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396762 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-slash\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396780 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-env-overrides\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396800 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovn-node-metrics-cert\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396821 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovnkube-script-lib\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396842 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-openvswitch\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396863 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-ovn\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396884 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8ccw\" (UniqueName: \"kubernetes.io/projected/087077d3-f9fa-4d16-aa7c-eb4241cbb293-kube-api-access-j8ccw\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396914 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-run-netns\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.396952 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-cni-bin\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: E1204 03:55:06.399725 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.404723 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.404769 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.404782 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.404799 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.404810 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:06Z","lastTransitionTime":"2025-12-04T03:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.409118 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: E1204 03:55:06.427459 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.427770 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.434034 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.434069 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.434078 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.434090 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.434100 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:06Z","lastTransitionTime":"2025-12-04T03:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.445939 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: E1204 03:55:06.446423 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.452290 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.452318 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.452326 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.452342 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.452351 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:06Z","lastTransitionTime":"2025-12-04T03:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.463103 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: E1204 03:55:06.480024 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.486330 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.486614 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.486631 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.486639 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.486652 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.486661 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:06Z","lastTransitionTime":"2025-12-04T03:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.497359 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-slash\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.497393 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-env-overrides\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.497409 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovn-node-metrics-cert\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.497425 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovnkube-script-lib\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.497441 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-openvswitch\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.497513 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-slash\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498007 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-ovn\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498041 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8ccw\" (UniqueName: \"kubernetes.io/projected/087077d3-f9fa-4d16-aa7c-eb4241cbb293-kube-api-access-j8ccw\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498378 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-run-netns\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498395 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-cni-bin\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498400 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-env-overrides\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498434 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-systemd\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498454 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-run-netns\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498484 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-cni-netd\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498494 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-cni-bin\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498537 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-systemd\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498456 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-cni-netd\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498580 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-kubelet\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498599 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498622 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-systemd-units\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498633 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovnkube-script-lib\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498663 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-etc-openvswitch\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498771 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-kubelet\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498772 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-ovn\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498820 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498847 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-systemd-units\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498896 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-openvswitch\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498939 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-etc-openvswitch\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.498979 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-node-log\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.499010 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-log-socket\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.499064 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovnkube-config\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.499088 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-run-ovn-kubernetes\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.499109 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-var-lib-openvswitch\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.499193 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-var-lib-openvswitch\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.499221 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-node-log\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.499255 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-log-socket\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.499693 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovnkube-config\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.499733 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-run-ovn-kubernetes\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.505856 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovn-node-metrics-cert\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.507805 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: E1204 03:55:06.508198 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: E1204 03:55:06.508357 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.513375 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.513410 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.513420 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.513432 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.513441 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:06Z","lastTransitionTime":"2025-12-04T03:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.522411 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-qdpsj" event={"ID":"5a576e68-334a-4a4f-aae0-888b6293dc41","Type":"ContainerStarted","Data":"583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.523812 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dsclk" event={"ID":"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577","Type":"ContainerStarted","Data":"49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.523843 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dsclk" event={"ID":"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577","Type":"ContainerStarted","Data":"85045d70fbe69286388ffd26bcde03943cd51cad7ea143e116359b2cf58c5acb"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.524797 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8ccw\" (UniqueName: \"kubernetes.io/projected/087077d3-f9fa-4d16-aa7c-eb4241cbb293-kube-api-access-j8ccw\") pod \"ovnkube-node-ljxsg\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.526356 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.527656 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hqps5" event={"ID":"8a1fde59-696b-48ad-8fce-491ab2ee8454","Type":"ContainerStarted","Data":"f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.527693 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hqps5" event={"ID":"8a1fde59-696b-48ad-8fce-491ab2ee8454","Type":"ContainerStarted","Data":"ed835111d73bc556e5fd76bcd8558358b35988df718919c772838ee6fb9870ca"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.530998 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.532207 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" event={"ID":"10aef7a5-56dc-4742-87d3-9c87074eb77a","Type":"ContainerStarted","Data":"02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.532286 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" event={"ID":"10aef7a5-56dc-4742-87d3-9c87074eb77a","Type":"ContainerStarted","Data":"f37255ebcdac7db8d61577dc63715963729cc6b93dd3eab65c0f1eb7da8efd64"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.535592 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.535639 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.535649 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"bbb9b8a48c0d140661a074a51a35137fffe28cf94a51d78e733f853960f411e1"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.543531 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.555617 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.566239 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.580113 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.594368 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.612236 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.615670 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.615700 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.615710 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.615723 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.615731 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:06Z","lastTransitionTime":"2025-12-04T03:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.628347 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.633879 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.645658 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: W1204 03:55:06.655392 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod087077d3_f9fa_4d16_aa7c_eb4241cbb293.slice/crio-0809ad07fd8b567e1c56cc8cea97f243e1f77063024a08be9619287825b8f072 WatchSource:0}: Error finding container 0809ad07fd8b567e1c56cc8cea97f243e1f77063024a08be9619287825b8f072: Status 404 returned error can't find the container with id 0809ad07fd8b567e1c56cc8cea97f243e1f77063024a08be9619287825b8f072 Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.669283 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.681759 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.692235 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.712470 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.718139 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.718164 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.718172 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.718184 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.718192 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:06Z","lastTransitionTime":"2025-12-04T03:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.726219 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.760354 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.818238 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.820019 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.820044 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.820052 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.820064 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.820075 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:06Z","lastTransitionTime":"2025-12-04T03:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.835745 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.882551 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.922339 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:06Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.922430 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.922451 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.922464 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.922480 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:06 crc kubenswrapper[4806]: I1204 03:55:06.922491 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:06Z","lastTransitionTime":"2025-12-04T03:55:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.004701 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.004741 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:07 crc kubenswrapper[4806]: E1204 03:55:07.004858 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 03:55:07 crc kubenswrapper[4806]: E1204 03:55:07.004881 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 03:55:07 crc kubenswrapper[4806]: E1204 03:55:07.004894 4806 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:07 crc kubenswrapper[4806]: E1204 03:55:07.004956 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:11.004939631 +0000 UTC m=+25.863452559 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:07 crc kubenswrapper[4806]: E1204 03:55:07.004857 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 03:55:07 crc kubenswrapper[4806]: E1204 03:55:07.004979 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 03:55:07 crc kubenswrapper[4806]: E1204 03:55:07.004986 4806 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:07 crc kubenswrapper[4806]: E1204 03:55:07.005007 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:11.004999953 +0000 UTC m=+25.863512881 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.024138 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.024167 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.024174 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.024186 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.024194 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:07Z","lastTransitionTime":"2025-12-04T03:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.105548 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.105643 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:07 crc kubenswrapper[4806]: E1204 03:55:07.105705 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:55:11.105688772 +0000 UTC m=+25.964201700 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:55:07 crc kubenswrapper[4806]: E1204 03:55:07.105742 4806 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 03:55:07 crc kubenswrapper[4806]: E1204 03:55:07.105775 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:11.105767555 +0000 UTC m=+25.964280473 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 03:55:07 crc kubenswrapper[4806]: E1204 03:55:07.105784 4806 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 03:55:07 crc kubenswrapper[4806]: E1204 03:55:07.105814 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:11.105806476 +0000 UTC m=+25.964319404 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.105741 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.126089 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.126119 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.126128 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.126140 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.126148 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:07Z","lastTransitionTime":"2025-12-04T03:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.228088 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.228129 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.228140 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.228156 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.228167 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:07Z","lastTransitionTime":"2025-12-04T03:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.329752 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.329796 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.329805 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.329823 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.329838 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:07Z","lastTransitionTime":"2025-12-04T03:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.423527 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.423552 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:07 crc kubenswrapper[4806]: E1204 03:55:07.424695 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.425461 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:07 crc kubenswrapper[4806]: E1204 03:55:07.425708 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:07 crc kubenswrapper[4806]: E1204 03:55:07.425863 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.435121 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.435370 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.435435 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.435495 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.435551 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:07Z","lastTransitionTime":"2025-12-04T03:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.538100 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.538444 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.538455 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.538469 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.538480 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:07Z","lastTransitionTime":"2025-12-04T03:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.539564 4806 generic.go:334] "Generic (PLEG): container finished" podID="10aef7a5-56dc-4742-87d3-9c87074eb77a" containerID="02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d" exitCode=0 Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.539641 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" event={"ID":"10aef7a5-56dc-4742-87d3-9c87074eb77a","Type":"ContainerDied","Data":"02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d"} Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.541395 4806 generic.go:334] "Generic (PLEG): container finished" podID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerID="413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a" exitCode=0 Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.541741 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerDied","Data":"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a"} Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.541763 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerStarted","Data":"0809ad07fd8b567e1c56cc8cea97f243e1f77063024a08be9619287825b8f072"} Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.573686 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.624812 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.637266 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.654195 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.654245 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.654254 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.654268 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.654277 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:07Z","lastTransitionTime":"2025-12-04T03:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.665216 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.683013 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.711303 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.731947 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.744803 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.756122 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.756254 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.756313 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.756370 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.756570 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:07Z","lastTransitionTime":"2025-12-04T03:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.760436 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.770364 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.785126 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.804840 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.821484 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.836848 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.854372 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.858314 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.858362 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.858375 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.858394 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.858406 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:07Z","lastTransitionTime":"2025-12-04T03:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.874501 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.891076 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.912435 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.933486 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.945759 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.960453 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.960501 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.960524 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.960554 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.960564 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:07Z","lastTransitionTime":"2025-12-04T03:55:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.969140 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.984170 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:07 crc kubenswrapper[4806]: I1204 03:55:07.995620 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:07Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.015305 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.030100 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.046632 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.057901 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.062837 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.062861 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.062870 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.062882 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.062892 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:08Z","lastTransitionTime":"2025-12-04T03:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.072269 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.165465 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.165503 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.165513 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.165525 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.165535 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:08Z","lastTransitionTime":"2025-12-04T03:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.267733 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.267769 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.267777 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.267791 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.267799 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:08Z","lastTransitionTime":"2025-12-04T03:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.370375 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.370416 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.370427 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.370441 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.370452 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:08Z","lastTransitionTime":"2025-12-04T03:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.473123 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.473178 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.473188 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.473206 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.473220 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:08Z","lastTransitionTime":"2025-12-04T03:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.546227 4806 generic.go:334] "Generic (PLEG): container finished" podID="10aef7a5-56dc-4742-87d3-9c87074eb77a" containerID="a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7" exitCode=0 Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.546302 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" event={"ID":"10aef7a5-56dc-4742-87d3-9c87074eb77a","Type":"ContainerDied","Data":"a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7"} Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.550000 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerStarted","Data":"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7"} Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.550038 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerStarted","Data":"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7"} Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.550047 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerStarted","Data":"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c"} Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.550055 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerStarted","Data":"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b"} Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.550063 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerStarted","Data":"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526"} Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.550071 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerStarted","Data":"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28"} Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.564622 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.577010 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.582951 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.582987 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.582998 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.583012 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.583020 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:08Z","lastTransitionTime":"2025-12-04T03:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.597858 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.624294 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.644347 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.657620 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.673991 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.685142 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.685163 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.685171 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.685183 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.685192 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:08Z","lastTransitionTime":"2025-12-04T03:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.691815 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.707764 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.722504 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.735539 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.747834 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.759548 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.787080 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.787113 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.787123 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.787135 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.787145 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:08Z","lastTransitionTime":"2025-12-04T03:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.787644 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.889265 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.889297 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.889305 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.889318 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.889327 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:08Z","lastTransitionTime":"2025-12-04T03:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.991328 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.991363 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.991375 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.991390 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:08 crc kubenswrapper[4806]: I1204 03:55:08.991402 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:08Z","lastTransitionTime":"2025-12-04T03:55:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.093567 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.094255 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.094271 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.094287 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.094297 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:09Z","lastTransitionTime":"2025-12-04T03:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.196503 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.196556 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.196571 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.196592 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.196607 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:09Z","lastTransitionTime":"2025-12-04T03:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.298497 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.298534 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.298543 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.298556 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.298565 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:09Z","lastTransitionTime":"2025-12-04T03:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.401226 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.401272 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.401285 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.401305 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.401320 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:09Z","lastTransitionTime":"2025-12-04T03:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.424803 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:09 crc kubenswrapper[4806]: E1204 03:55:09.424937 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.425230 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:09 crc kubenswrapper[4806]: E1204 03:55:09.425281 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.425494 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:09 crc kubenswrapper[4806]: E1204 03:55:09.425758 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.504233 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.504578 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.504681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.504793 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.504887 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:09Z","lastTransitionTime":"2025-12-04T03:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.555502 4806 generic.go:334] "Generic (PLEG): container finished" podID="10aef7a5-56dc-4742-87d3-9c87074eb77a" containerID="dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f" exitCode=0 Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.555542 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" event={"ID":"10aef7a5-56dc-4742-87d3-9c87074eb77a","Type":"ContainerDied","Data":"dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f"} Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.573686 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:09Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.587314 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:09Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.598104 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:09Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.615905 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.616302 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.616488 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.616637 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.616765 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:09Z","lastTransitionTime":"2025-12-04T03:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.633689 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:09Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.661706 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:09Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.686979 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:09Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.701159 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:09Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.714854 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:09Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.722004 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.722041 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.722051 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.722062 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.722072 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:09Z","lastTransitionTime":"2025-12-04T03:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.728806 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:09Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.744641 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:09Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.757325 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:09Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.769729 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:09Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.781766 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:09Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.791569 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:09Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.823908 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.823971 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.823983 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.824000 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.824012 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:09Z","lastTransitionTime":"2025-12-04T03:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.927228 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.927279 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.927291 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.927308 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:09 crc kubenswrapper[4806]: I1204 03:55:09.927321 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:09Z","lastTransitionTime":"2025-12-04T03:55:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.031187 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.031271 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.031294 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.031324 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.031345 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:10Z","lastTransitionTime":"2025-12-04T03:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.133238 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.133276 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.133284 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.133296 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.133309 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:10Z","lastTransitionTime":"2025-12-04T03:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.235445 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.235498 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.235515 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.235533 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.235547 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:10Z","lastTransitionTime":"2025-12-04T03:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.338596 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.338637 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.338648 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.338663 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.338676 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:10Z","lastTransitionTime":"2025-12-04T03:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.440902 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.440950 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.440959 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.440971 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.440979 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:10Z","lastTransitionTime":"2025-12-04T03:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.543578 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.543654 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.543675 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.543699 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.543716 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:10Z","lastTransitionTime":"2025-12-04T03:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.566194 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerStarted","Data":"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546"} Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.569456 4806 generic.go:334] "Generic (PLEG): container finished" podID="10aef7a5-56dc-4742-87d3-9c87074eb77a" containerID="8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32" exitCode=0 Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.569508 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" event={"ID":"10aef7a5-56dc-4742-87d3-9c87074eb77a","Type":"ContainerDied","Data":"8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32"} Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.603665 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:10Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.623559 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:10Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.636185 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:10Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.648710 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.648779 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.648801 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.648830 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.648853 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:10Z","lastTransitionTime":"2025-12-04T03:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.655198 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:10Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.670028 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:10Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.696990 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:10Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.711250 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:10Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.724750 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:10Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.740085 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:10Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.753852 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:10Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.754980 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.755012 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.755022 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.755039 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.755050 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:10Z","lastTransitionTime":"2025-12-04T03:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.772173 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:10Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.786611 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:10Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.800403 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:10Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.816677 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:10Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.858299 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.858580 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.858588 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.858600 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.858609 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:10Z","lastTransitionTime":"2025-12-04T03:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.961068 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.961274 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.961354 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.961414 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:10 crc kubenswrapper[4806]: I1204 03:55:10.961696 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:10Z","lastTransitionTime":"2025-12-04T03:55:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.048539 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.048976 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:11 crc kubenswrapper[4806]: E1204 03:55:11.048894 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 03:55:11 crc kubenswrapper[4806]: E1204 03:55:11.049280 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 03:55:11 crc kubenswrapper[4806]: E1204 03:55:11.049370 4806 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:11 crc kubenswrapper[4806]: E1204 03:55:11.049508 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:19.049489636 +0000 UTC m=+33.908002564 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:11 crc kubenswrapper[4806]: E1204 03:55:11.049182 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 03:55:11 crc kubenswrapper[4806]: E1204 03:55:11.050087 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 03:55:11 crc kubenswrapper[4806]: E1204 03:55:11.050167 4806 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:11 crc kubenswrapper[4806]: E1204 03:55:11.050271 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:19.05025988 +0000 UTC m=+33.908772808 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.063897 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.063953 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.063963 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.063977 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.063988 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:11Z","lastTransitionTime":"2025-12-04T03:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.149872 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.150032 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:11 crc kubenswrapper[4806]: E1204 03:55:11.150135 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:55:19.150103633 +0000 UTC m=+34.008616561 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:55:11 crc kubenswrapper[4806]: E1204 03:55:11.150158 4806 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 03:55:11 crc kubenswrapper[4806]: E1204 03:55:11.150222 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:19.150204096 +0000 UTC m=+34.008717034 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.150212 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:11 crc kubenswrapper[4806]: E1204 03:55:11.150287 4806 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 03:55:11 crc kubenswrapper[4806]: E1204 03:55:11.150329 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:19.15032028 +0000 UTC m=+34.008833318 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.166216 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.166438 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.166609 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.166752 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.166817 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:11Z","lastTransitionTime":"2025-12-04T03:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.269164 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.269198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.269205 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.269218 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.269228 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:11Z","lastTransitionTime":"2025-12-04T03:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.373253 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.373285 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.373294 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.373306 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.373314 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:11Z","lastTransitionTime":"2025-12-04T03:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.423812 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.423840 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.424125 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:11 crc kubenswrapper[4806]: E1204 03:55:11.424236 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:11 crc kubenswrapper[4806]: E1204 03:55:11.424085 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:11 crc kubenswrapper[4806]: E1204 03:55:11.424315 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.476011 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.476245 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.476322 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.476552 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.476745 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:11Z","lastTransitionTime":"2025-12-04T03:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.574738 4806 generic.go:334] "Generic (PLEG): container finished" podID="10aef7a5-56dc-4742-87d3-9c87074eb77a" containerID="d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429" exitCode=0 Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.574789 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" event={"ID":"10aef7a5-56dc-4742-87d3-9c87074eb77a","Type":"ContainerDied","Data":"d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429"} Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.583303 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.583363 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.583381 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.583406 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.583424 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:11Z","lastTransitionTime":"2025-12-04T03:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.586673 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:11Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.610112 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:11Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.623797 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:11Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.636265 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:11Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.648144 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:11Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.662575 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:11Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.673107 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:11Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.686012 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.686049 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.686062 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.686078 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.686090 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:11Z","lastTransitionTime":"2025-12-04T03:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.686864 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:11Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.702679 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:11Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.713724 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:11Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.727463 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:11Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.741088 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:11Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.751501 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:11Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.764636 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:11Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.788855 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.788913 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.788954 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.788978 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.788995 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:11Z","lastTransitionTime":"2025-12-04T03:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.894309 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.894386 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.894404 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.894428 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.894447 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:11Z","lastTransitionTime":"2025-12-04T03:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.997832 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.997964 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.997983 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.998007 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:11 crc kubenswrapper[4806]: I1204 03:55:11.998024 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:11Z","lastTransitionTime":"2025-12-04T03:55:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.101408 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.101456 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.101474 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.101496 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.101513 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:12Z","lastTransitionTime":"2025-12-04T03:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.204726 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.204757 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.204767 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.204782 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.204792 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:12Z","lastTransitionTime":"2025-12-04T03:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.307235 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.307277 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.307290 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.307306 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.307318 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:12Z","lastTransitionTime":"2025-12-04T03:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.411896 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.412429 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.412442 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.412461 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.412472 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:12Z","lastTransitionTime":"2025-12-04T03:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.514612 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.514657 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.514669 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.514685 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.514697 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:12Z","lastTransitionTime":"2025-12-04T03:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.583227 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerStarted","Data":"263099c30f9fffa3519185b3e1cd0aebbd8dfb117ffdcdef6d3175551771aa2b"} Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.583539 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.589437 4806 generic.go:334] "Generic (PLEG): container finished" podID="10aef7a5-56dc-4742-87d3-9c87074eb77a" containerID="2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7" exitCode=0 Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.589484 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" event={"ID":"10aef7a5-56dc-4742-87d3-9c87074eb77a","Type":"ContainerDied","Data":"2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7"} Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.606010 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.616876 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.616913 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.616936 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.616950 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.616959 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:12Z","lastTransitionTime":"2025-12-04T03:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.618673 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.619296 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.628070 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.644039 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://263099c30f9fffa3519185b3e1cd0aebbd8dfb117ffdcdef6d3175551771aa2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.661791 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.676190 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.686701 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.698145 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.712489 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.719593 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.719623 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.719632 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.719646 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.719656 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:12Z","lastTransitionTime":"2025-12-04T03:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.723804 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.739319 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.752272 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.767820 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.779525 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.800612 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://263099c30f9fffa3519185b3e1cd0aebbd8dfb117ffdcdef6d3175551771aa2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.813129 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.821817 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.821857 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.821869 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.821885 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.821896 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:12Z","lastTransitionTime":"2025-12-04T03:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.824146 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.833451 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.844510 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.857383 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.867631 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.881646 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.897006 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.909884 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.924469 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.924518 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.924529 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.924542 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.924553 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:12Z","lastTransitionTime":"2025-12-04T03:55:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.927793 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.938875 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.952573 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:12 crc kubenswrapper[4806]: I1204 03:55:12.967939 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:12Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.026943 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.026980 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.026988 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.027018 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.027029 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:13Z","lastTransitionTime":"2025-12-04T03:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.129114 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.129150 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.129158 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.129172 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.129180 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:13Z","lastTransitionTime":"2025-12-04T03:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.231456 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.231504 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.231518 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.231535 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.231546 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:13Z","lastTransitionTime":"2025-12-04T03:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.334168 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.334203 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.334211 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.334225 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.334234 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:13Z","lastTransitionTime":"2025-12-04T03:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.422640 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.422754 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.422689 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:13 crc kubenswrapper[4806]: E1204 03:55:13.422986 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:13 crc kubenswrapper[4806]: E1204 03:55:13.423034 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:13 crc kubenswrapper[4806]: E1204 03:55:13.423101 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.436214 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.436257 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.436270 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.436287 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.436300 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:13Z","lastTransitionTime":"2025-12-04T03:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.538837 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.538882 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.538894 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.538910 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.538949 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:13Z","lastTransitionTime":"2025-12-04T03:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.598894 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" event={"ID":"10aef7a5-56dc-4742-87d3-9c87074eb77a","Type":"ContainerStarted","Data":"09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996"} Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.599012 4806 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.599745 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.616296 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.624948 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.630730 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.641194 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.641376 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.641527 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.641701 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.641825 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:13Z","lastTransitionTime":"2025-12-04T03:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.643734 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.667128 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://263099c30f9fffa3519185b3e1cd0aebbd8dfb117ffdcdef6d3175551771aa2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.681731 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.693225 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.706724 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.718019 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.731961 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.743950 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.743982 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.744104 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.744115 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.744130 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.744142 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:13Z","lastTransitionTime":"2025-12-04T03:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.756654 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.769406 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.781022 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.793384 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.804894 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.815603 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.829437 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.841260 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.846972 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.847004 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.847015 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.847031 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.847043 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:13Z","lastTransitionTime":"2025-12-04T03:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.852387 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.866525 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.877617 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.886543 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.905136 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://263099c30f9fffa3519185b3e1cd0aebbd8dfb117ffdcdef6d3175551771aa2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.917180 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.932115 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.944062 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.949959 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.950009 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.950022 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.950040 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.950057 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:13Z","lastTransitionTime":"2025-12-04T03:55:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.959782 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:13 crc kubenswrapper[4806]: I1204 03:55:13.971559 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:13Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.052387 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.052680 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.052692 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.052706 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.052718 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:14Z","lastTransitionTime":"2025-12-04T03:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.156137 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.156178 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.156322 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.156339 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.156349 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:14Z","lastTransitionTime":"2025-12-04T03:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.259489 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.259542 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.259555 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.259575 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.259588 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:14Z","lastTransitionTime":"2025-12-04T03:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.361861 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.361890 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.361897 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.361909 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.361918 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:14Z","lastTransitionTime":"2025-12-04T03:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.464253 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.464294 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.464306 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.464323 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.464333 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:14Z","lastTransitionTime":"2025-12-04T03:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.566451 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.566483 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.566491 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.566510 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.566520 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:14Z","lastTransitionTime":"2025-12-04T03:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.601176 4806 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.668779 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.669097 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.669233 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.669384 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.669555 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:14Z","lastTransitionTime":"2025-12-04T03:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.771823 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.771859 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.771869 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.771885 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.771896 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:14Z","lastTransitionTime":"2025-12-04T03:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.874495 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.874536 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.874545 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.874557 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.874566 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:14Z","lastTransitionTime":"2025-12-04T03:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.979368 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.979419 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.979439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.980146 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:14 crc kubenswrapper[4806]: I1204 03:55:14.980273 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:14Z","lastTransitionTime":"2025-12-04T03:55:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.082786 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.082822 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.082833 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.082847 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.082859 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:15Z","lastTransitionTime":"2025-12-04T03:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.185814 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.185855 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.185864 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.185878 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.185890 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:15Z","lastTransitionTime":"2025-12-04T03:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.287768 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.287819 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.287832 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.287848 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.287859 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:15Z","lastTransitionTime":"2025-12-04T03:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.390330 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.390371 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.390383 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.390398 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.390409 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:15Z","lastTransitionTime":"2025-12-04T03:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.422982 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.423018 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.423174 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:15 crc kubenswrapper[4806]: E1204 03:55:15.423174 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:15 crc kubenswrapper[4806]: E1204 03:55:15.423347 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:15 crc kubenswrapper[4806]: E1204 03:55:15.423455 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.444975 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.466102 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.481060 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.492691 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.492762 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.492779 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.492798 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.492810 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:15Z","lastTransitionTime":"2025-12-04T03:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.496479 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.513164 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.525488 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.569975 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://263099c30f9fffa3519185b3e1cd0aebbd8dfb117ffdcdef6d3175551771aa2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.589229 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.594863 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.594906 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.594917 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.594946 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.594963 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:15Z","lastTransitionTime":"2025-12-04T03:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.605596 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/0.log" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.607530 4806 generic.go:334] "Generic (PLEG): container finished" podID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerID="263099c30f9fffa3519185b3e1cd0aebbd8dfb117ffdcdef6d3175551771aa2b" exitCode=1 Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.607606 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerDied","Data":"263099c30f9fffa3519185b3e1cd0aebbd8dfb117ffdcdef6d3175551771aa2b"} Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.608129 4806 scope.go:117] "RemoveContainer" containerID="263099c30f9fffa3519185b3e1cd0aebbd8dfb117ffdcdef6d3175551771aa2b" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.616392 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.635900 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.650086 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.668296 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.682384 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.693761 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.697301 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.697344 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.697354 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.697369 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.697381 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:15Z","lastTransitionTime":"2025-12-04T03:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.705996 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.718974 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.739060 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.750637 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.763329 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.773003 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.784538 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.793977 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.800038 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.800142 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.800154 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.800170 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.800180 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:15Z","lastTransitionTime":"2025-12-04T03:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.810674 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://263099c30f9fffa3519185b3e1cd0aebbd8dfb117ffdcdef6d3175551771aa2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://263099c30f9fffa3519185b3e1cd0aebbd8dfb117ffdcdef6d3175551771aa2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:14Z\\\",\\\"message\\\":\\\" 03:55:14.530277 5972 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 03:55:14.530294 5972 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 03:55:14.530305 5972 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 03:55:14.530309 5972 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 03:55:14.530294 5972 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 03:55:14.530325 5972 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1204 03:55:14.530330 5972 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1204 03:55:14.530333 5972 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 03:55:14.530338 5972 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 03:55:14.530346 5972 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 03:55:14.530354 5972 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 03:55:14.530355 5972 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 03:55:14.530355 5972 factory.go:656] Stopping watch factory\\\\nI1204 03:55:14.530364 5972 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 03:55:14.530378 5972 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 03:55:14.530324 5972 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.823583 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.837313 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.849148 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.860919 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.873917 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.902762 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.902803 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.902815 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.902831 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:15 crc kubenswrapper[4806]: I1204 03:55:15.902842 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:15Z","lastTransitionTime":"2025-12-04T03:55:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.005029 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.005070 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.005079 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.005093 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.005102 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:16Z","lastTransitionTime":"2025-12-04T03:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.107817 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.107859 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.107868 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.107882 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.107893 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:16Z","lastTransitionTime":"2025-12-04T03:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.210822 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.210867 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.210878 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.210899 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.210911 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:16Z","lastTransitionTime":"2025-12-04T03:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.312871 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.312942 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.312955 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.312969 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.312980 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:16Z","lastTransitionTime":"2025-12-04T03:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.415626 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.415681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.415690 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.415703 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.415711 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:16Z","lastTransitionTime":"2025-12-04T03:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.518020 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.518046 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.518053 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.518066 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.518078 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:16Z","lastTransitionTime":"2025-12-04T03:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.615456 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/0.log" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.619666 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.619702 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.619715 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.619731 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.619742 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:16Z","lastTransitionTime":"2025-12-04T03:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.620306 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerStarted","Data":"1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297"} Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.620367 4806 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.643240 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.658018 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.670724 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.689632 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://263099c30f9fffa3519185b3e1cd0aebbd8dfb117ffdcdef6d3175551771aa2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:14Z\\\",\\\"message\\\":\\\" 03:55:14.530277 5972 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 03:55:14.530294 5972 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 03:55:14.530305 5972 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 03:55:14.530309 5972 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 03:55:14.530294 5972 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 03:55:14.530325 5972 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1204 03:55:14.530330 5972 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1204 03:55:14.530333 5972 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 03:55:14.530338 5972 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 03:55:14.530346 5972 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 03:55:14.530354 5972 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 03:55:14.530355 5972 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 03:55:14.530355 5972 factory.go:656] Stopping watch factory\\\\nI1204 03:55:14.530364 5972 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 03:55:14.530378 5972 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 03:55:14.530324 5972 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.703538 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.718223 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.722844 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.722882 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.722894 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.722917 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.722945 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:16Z","lastTransitionTime":"2025-12-04T03:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.730284 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.742256 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.746437 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.746491 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.746507 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.746532 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.746551 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:16Z","lastTransitionTime":"2025-12-04T03:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.753316 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: E1204 03:55:16.760523 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.764566 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.764632 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.764650 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.764673 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.764691 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:16Z","lastTransitionTime":"2025-12-04T03:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.764697 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: E1204 03:55:16.779698 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.780560 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.784349 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.784912 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.784944 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.784961 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.784972 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:16Z","lastTransitionTime":"2025-12-04T03:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.797905 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: E1204 03:55:16.802030 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.807002 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.807058 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.807070 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.807092 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.807106 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:16Z","lastTransitionTime":"2025-12-04T03:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.813520 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: E1204 03:55:16.821136 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.825232 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.825695 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.825858 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.826118 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.826324 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.826424 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:16Z","lastTransitionTime":"2025-12-04T03:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:16 crc kubenswrapper[4806]: E1204 03:55:16.840736 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:16Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:16 crc kubenswrapper[4806]: E1204 03:55:16.840858 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.842538 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.842572 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.842582 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.842596 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.842606 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:16Z","lastTransitionTime":"2025-12-04T03:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.945668 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.945707 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.945721 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.945735 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:16 crc kubenswrapper[4806]: I1204 03:55:16.945747 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:16Z","lastTransitionTime":"2025-12-04T03:55:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.048731 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.048776 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.048786 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.048801 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.048812 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:17Z","lastTransitionTime":"2025-12-04T03:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.151636 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.151683 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.151697 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.151748 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.151766 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:17Z","lastTransitionTime":"2025-12-04T03:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.255625 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.256064 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.256247 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.256484 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.256676 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:17Z","lastTransitionTime":"2025-12-04T03:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.359672 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.359722 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.359741 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.359762 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.359779 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:17Z","lastTransitionTime":"2025-12-04T03:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.423052 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.423167 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:17 crc kubenswrapper[4806]: E1204 03:55:17.423261 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.423331 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:17 crc kubenswrapper[4806]: E1204 03:55:17.423453 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:17 crc kubenswrapper[4806]: E1204 03:55:17.423526 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.462290 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.462353 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.462379 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.462409 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.462436 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:17Z","lastTransitionTime":"2025-12-04T03:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.565414 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.565782 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.565915 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.566135 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.566282 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:17Z","lastTransitionTime":"2025-12-04T03:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.625043 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/1.log" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.626090 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/0.log" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.629290 4806 generic.go:334] "Generic (PLEG): container finished" podID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerID="1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297" exitCode=1 Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.629339 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerDied","Data":"1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297"} Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.629398 4806 scope.go:117] "RemoveContainer" containerID="263099c30f9fffa3519185b3e1cd0aebbd8dfb117ffdcdef6d3175551771aa2b" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.633780 4806 scope.go:117] "RemoveContainer" containerID="1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297" Dec 04 03:55:17 crc kubenswrapper[4806]: E1204 03:55:17.634335 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.660018 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:17Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.668044 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.668109 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.668125 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.668146 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.668161 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:17Z","lastTransitionTime":"2025-12-04T03:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.677321 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:17Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.689915 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:17Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.713138 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://263099c30f9fffa3519185b3e1cd0aebbd8dfb117ffdcdef6d3175551771aa2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:14Z\\\",\\\"message\\\":\\\" 03:55:14.530277 5972 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 03:55:14.530294 5972 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 03:55:14.530305 5972 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 03:55:14.530309 5972 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 03:55:14.530294 5972 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 03:55:14.530325 5972 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1204 03:55:14.530330 5972 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1204 03:55:14.530333 5972 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 03:55:14.530338 5972 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 03:55:14.530346 5972 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 03:55:14.530354 5972 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 03:55:14.530355 5972 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 03:55:14.530355 5972 factory.go:656] Stopping watch factory\\\\nI1204 03:55:14.530364 5972 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 03:55:14.530378 5972 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 03:55:14.530324 5972 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\":55:16.345209 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 03:55:16.345217 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 03:55:16.345224 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 03:55:16.345231 6103 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 03:55:16.345302 6103 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345448 6103 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 03:55:16.345527 6103 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345754 6103 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345963 6103 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346267 6103 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346350 6103 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:17Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.736520 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:17Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.754666 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:17Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.770234 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:17Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.771068 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.771157 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.771182 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.771225 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.771248 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:17Z","lastTransitionTime":"2025-12-04T03:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.783303 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:17Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.799590 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:17Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.813988 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:17Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.829320 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:17Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.846701 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:17Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.859502 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:17Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.869820 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:17Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.873706 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.873738 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.873745 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.873759 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.873768 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:17Z","lastTransitionTime":"2025-12-04T03:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.975727 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.975785 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.975804 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.975826 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:17 crc kubenswrapper[4806]: I1204 03:55:17.975840 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:17Z","lastTransitionTime":"2025-12-04T03:55:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.079112 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.079166 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.079184 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.079206 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.079226 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:18Z","lastTransitionTime":"2025-12-04T03:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.182036 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.182392 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.182547 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.182660 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.182760 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:18Z","lastTransitionTime":"2025-12-04T03:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.224015 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd"] Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.224785 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.230195 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.230455 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.248547 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.268676 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.285397 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.285430 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.285438 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.285452 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.285460 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:18Z","lastTransitionTime":"2025-12-04T03:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.287345 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.308894 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.324700 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.333840 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5d7e1e32-7021-43cb-a282-653d6c4dd750-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lgrcd\" (UID: \"5d7e1e32-7021-43cb-a282-653d6c4dd750\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.333888 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5d7e1e32-7021-43cb-a282-653d6c4dd750-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lgrcd\" (UID: \"5d7e1e32-7021-43cb-a282-653d6c4dd750\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.333962 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5d7e1e32-7021-43cb-a282-653d6c4dd750-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lgrcd\" (UID: \"5d7e1e32-7021-43cb-a282-653d6c4dd750\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.334007 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f62m9\" (UniqueName: \"kubernetes.io/projected/5d7e1e32-7021-43cb-a282-653d6c4dd750-kube-api-access-f62m9\") pod \"ovnkube-control-plane-749d76644c-lgrcd\" (UID: \"5d7e1e32-7021-43cb-a282-653d6c4dd750\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.336722 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.357261 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.370648 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.381052 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.388525 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.388600 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.388613 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.388632 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.388651 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:18Z","lastTransitionTime":"2025-12-04T03:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.395854 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.407874 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.428894 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://263099c30f9fffa3519185b3e1cd0aebbd8dfb117ffdcdef6d3175551771aa2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:14Z\\\",\\\"message\\\":\\\" 03:55:14.530277 5972 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 03:55:14.530294 5972 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1204 03:55:14.530305 5972 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1204 03:55:14.530309 5972 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1204 03:55:14.530294 5972 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 03:55:14.530325 5972 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1204 03:55:14.530330 5972 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1204 03:55:14.530333 5972 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 03:55:14.530338 5972 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 03:55:14.530346 5972 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 03:55:14.530354 5972 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 03:55:14.530355 5972 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1204 03:55:14.530355 5972 factory.go:656] Stopping watch factory\\\\nI1204 03:55:14.530364 5972 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1204 03:55:14.530378 5972 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1204 03:55:14.530324 5972 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\":55:16.345209 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 03:55:16.345217 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 03:55:16.345224 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 03:55:16.345231 6103 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 03:55:16.345302 6103 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345448 6103 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 03:55:16.345527 6103 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345754 6103 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345963 6103 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346267 6103 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346350 6103 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.434978 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5d7e1e32-7021-43cb-a282-653d6c4dd750-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lgrcd\" (UID: \"5d7e1e32-7021-43cb-a282-653d6c4dd750\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.435099 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f62m9\" (UniqueName: \"kubernetes.io/projected/5d7e1e32-7021-43cb-a282-653d6c4dd750-kube-api-access-f62m9\") pod \"ovnkube-control-plane-749d76644c-lgrcd\" (UID: \"5d7e1e32-7021-43cb-a282-653d6c4dd750\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.435139 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5d7e1e32-7021-43cb-a282-653d6c4dd750-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lgrcd\" (UID: \"5d7e1e32-7021-43cb-a282-653d6c4dd750\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.435171 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5d7e1e32-7021-43cb-a282-653d6c4dd750-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lgrcd\" (UID: \"5d7e1e32-7021-43cb-a282-653d6c4dd750\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.435782 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5d7e1e32-7021-43cb-a282-653d6c4dd750-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-lgrcd\" (UID: \"5d7e1e32-7021-43cb-a282-653d6c4dd750\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.436036 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5d7e1e32-7021-43cb-a282-653d6c4dd750-env-overrides\") pod \"ovnkube-control-plane-749d76644c-lgrcd\" (UID: \"5d7e1e32-7021-43cb-a282-653d6c4dd750\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.441884 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5d7e1e32-7021-43cb-a282-653d6c4dd750-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-lgrcd\" (UID: \"5d7e1e32-7021-43cb-a282-653d6c4dd750\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.442915 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.453097 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.454133 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f62m9\" (UniqueName: \"kubernetes.io/projected/5d7e1e32-7021-43cb-a282-653d6c4dd750-kube-api-access-f62m9\") pod \"ovnkube-control-plane-749d76644c-lgrcd\" (UID: \"5d7e1e32-7021-43cb-a282-653d6c4dd750\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.463734 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.492081 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.492430 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.492534 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.492705 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.492840 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:18Z","lastTransitionTime":"2025-12-04T03:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.549048 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" Dec 04 03:55:18 crc kubenswrapper[4806]: W1204 03:55:18.564602 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d7e1e32_7021_43cb_a282_653d6c4dd750.slice/crio-e819810ff3edd70732673b08528426f383f163e806e6fe14180264ba1fcc890e WatchSource:0}: Error finding container e819810ff3edd70732673b08528426f383f163e806e6fe14180264ba1fcc890e: Status 404 returned error can't find the container with id e819810ff3edd70732673b08528426f383f163e806e6fe14180264ba1fcc890e Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.595115 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.595153 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.595167 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.595185 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.595196 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:18Z","lastTransitionTime":"2025-12-04T03:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.635572 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/1.log" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.645241 4806 scope.go:117] "RemoveContainer" containerID="1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297" Dec 04 03:55:18 crc kubenswrapper[4806]: E1204 03:55:18.645484 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.646272 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" event={"ID":"5d7e1e32-7021-43cb-a282-653d6c4dd750","Type":"ContainerStarted","Data":"e819810ff3edd70732673b08528426f383f163e806e6fe14180264ba1fcc890e"} Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.660291 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.671681 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.690067 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\":55:16.345209 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 03:55:16.345217 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 03:55:16.345224 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 03:55:16.345231 6103 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 03:55:16.345302 6103 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345448 6103 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 03:55:16.345527 6103 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345754 6103 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345963 6103 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346267 6103 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346350 6103 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.698526 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.698574 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.698593 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.698615 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.698632 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:18Z","lastTransitionTime":"2025-12-04T03:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.706822 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.720452 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.734791 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.747206 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.760467 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.777184 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.788393 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.801366 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.801999 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.802042 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.802055 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.802073 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.802086 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:18Z","lastTransitionTime":"2025-12-04T03:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.813116 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.829023 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.843441 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.858794 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.904348 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.904393 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.904402 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.904416 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:18 crc kubenswrapper[4806]: I1204 03:55:18.904427 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:18Z","lastTransitionTime":"2025-12-04T03:55:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.006326 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.006354 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.006363 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.006374 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.006385 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:19Z","lastTransitionTime":"2025-12-04T03:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.109071 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.109116 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.109129 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.109146 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.109157 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:19Z","lastTransitionTime":"2025-12-04T03:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.140549 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.140619 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.140707 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.140733 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.140744 4806 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.140760 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.140789 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.140805 4806 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.140793 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:35.140780047 +0000 UTC m=+49.999292975 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.140862 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:35.140846899 +0000 UTC m=+49.999359837 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.211185 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.211329 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.211352 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.211380 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.211407 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:19Z","lastTransitionTime":"2025-12-04T03:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.242098 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.242328 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:55:35.242293483 +0000 UTC m=+50.100806451 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.242395 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.242465 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.242580 4806 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.242620 4806 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.242660 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:35.242638684 +0000 UTC m=+50.101151652 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.242702 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:35.242679505 +0000 UTC m=+50.101192473 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.314581 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.314637 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.314655 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.314683 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.314706 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:19Z","lastTransitionTime":"2025-12-04T03:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.416987 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.417632 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.417729 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.417818 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.417909 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:19Z","lastTransitionTime":"2025-12-04T03:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.424181 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.424676 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.424857 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.425026 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.425191 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.425439 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.521456 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.521886 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.522023 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.522171 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.522293 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:19Z","lastTransitionTime":"2025-12-04T03:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.625449 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.625580 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.625596 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.625618 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.625636 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:19Z","lastTransitionTime":"2025-12-04T03:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.651548 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" event={"ID":"5d7e1e32-7021-43cb-a282-653d6c4dd750","Type":"ContainerStarted","Data":"554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066"} Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.651612 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" event={"ID":"5d7e1e32-7021-43cb-a282-653d6c4dd750","Type":"ContainerStarted","Data":"c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d"} Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.669668 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.687807 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.703483 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.714588 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-b9khd"] Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.715059 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.715119 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.720667 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.728322 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.728359 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.728370 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.728387 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.728397 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:19Z","lastTransitionTime":"2025-12-04T03:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.736131 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.751677 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.773300 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\":55:16.345209 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 03:55:16.345217 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 03:55:16.345224 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 03:55:16.345231 6103 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 03:55:16.345302 6103 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345448 6103 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 03:55:16.345527 6103 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345754 6103 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345963 6103 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346267 6103 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346350 6103 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.787845 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.801485 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.813735 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.827709 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.830237 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.830305 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.830330 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.830358 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.830380 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:19Z","lastTransitionTime":"2025-12-04T03:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.842385 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.849507 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs\") pod \"network-metrics-daemon-b9khd\" (UID: \"41fdb600-3868-48ab-8396-17c2382b6168\") " pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.849540 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd5dc\" (UniqueName: \"kubernetes.io/projected/41fdb600-3868-48ab-8396-17c2382b6168-kube-api-access-xd5dc\") pod \"network-metrics-daemon-b9khd\" (UID: \"41fdb600-3868-48ab-8396-17c2382b6168\") " pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.860104 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.875026 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.888433 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.899024 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.918318 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\":55:16.345209 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 03:55:16.345217 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 03:55:16.345224 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 03:55:16.345231 6103 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 03:55:16.345302 6103 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345448 6103 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 03:55:16.345527 6103 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345754 6103 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345963 6103 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346267 6103 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346350 6103 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.928518 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b9khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41fdb600-3868-48ab-8396-17c2382b6168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b9khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.932415 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.932444 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.932455 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.932472 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.932482 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:19Z","lastTransitionTime":"2025-12-04T03:55:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.940599 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.950450 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs\") pod \"network-metrics-daemon-b9khd\" (UID: \"41fdb600-3868-48ab-8396-17c2382b6168\") " pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.950503 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd5dc\" (UniqueName: \"kubernetes.io/projected/41fdb600-3868-48ab-8396-17c2382b6168-kube-api-access-xd5dc\") pod \"network-metrics-daemon-b9khd\" (UID: \"41fdb600-3868-48ab-8396-17c2382b6168\") " pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.950613 4806 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 03:55:19 crc kubenswrapper[4806]: E1204 03:55:19.950708 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs podName:41fdb600-3868-48ab-8396-17c2382b6168 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:20.450680221 +0000 UTC m=+35.309193179 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs") pod "network-metrics-daemon-b9khd" (UID: "41fdb600-3868-48ab-8396-17c2382b6168") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.952633 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.964034 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.970476 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd5dc\" (UniqueName: \"kubernetes.io/projected/41fdb600-3868-48ab-8396-17c2382b6168-kube-api-access-xd5dc\") pod \"network-metrics-daemon-b9khd\" (UID: \"41fdb600-3868-48ab-8396-17c2382b6168\") " pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.983769 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:19 crc kubenswrapper[4806]: I1204 03:55:19.998752 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:19Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.016759 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:20Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.030293 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:20Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.041357 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.041398 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.041410 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.041429 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.041442 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:20Z","lastTransitionTime":"2025-12-04T03:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.046458 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:20Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.064730 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:20Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.078258 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:20Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.091645 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:20Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.103178 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:20Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.120699 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:20Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.143195 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.143232 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.143247 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.143267 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.143283 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:20Z","lastTransitionTime":"2025-12-04T03:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.245837 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.245890 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.245902 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.245939 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.245953 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:20Z","lastTransitionTime":"2025-12-04T03:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.349141 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.349190 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.349201 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.349221 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.349232 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:20Z","lastTransitionTime":"2025-12-04T03:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.451732 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.451786 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.451801 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.451821 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.451835 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:20Z","lastTransitionTime":"2025-12-04T03:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.455354 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs\") pod \"network-metrics-daemon-b9khd\" (UID: \"41fdb600-3868-48ab-8396-17c2382b6168\") " pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:20 crc kubenswrapper[4806]: E1204 03:55:20.455490 4806 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 03:55:20 crc kubenswrapper[4806]: E1204 03:55:20.455556 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs podName:41fdb600-3868-48ab-8396-17c2382b6168 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:21.455540233 +0000 UTC m=+36.314053171 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs") pod "network-metrics-daemon-b9khd" (UID: "41fdb600-3868-48ab-8396-17c2382b6168") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.555263 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.555339 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.555365 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.555396 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.555418 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:20Z","lastTransitionTime":"2025-12-04T03:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.658271 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.658347 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.658373 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.658406 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.658429 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:20Z","lastTransitionTime":"2025-12-04T03:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.761467 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.761512 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.761523 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.761541 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.761551 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:20Z","lastTransitionTime":"2025-12-04T03:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.864405 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.864455 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.864465 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.864487 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.864498 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:20Z","lastTransitionTime":"2025-12-04T03:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.966525 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.966557 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.966565 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.966578 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:20 crc kubenswrapper[4806]: I1204 03:55:20.966586 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:20Z","lastTransitionTime":"2025-12-04T03:55:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.069726 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.070217 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.070373 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.070535 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.070662 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:21Z","lastTransitionTime":"2025-12-04T03:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.173325 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.173391 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.173409 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.173438 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.173456 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:21Z","lastTransitionTime":"2025-12-04T03:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.276423 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.276504 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.276529 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.276562 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.276585 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:21Z","lastTransitionTime":"2025-12-04T03:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.379992 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.380035 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.380048 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.380067 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.380082 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:21Z","lastTransitionTime":"2025-12-04T03:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.422914 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.423256 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.423290 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.423337 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:21 crc kubenswrapper[4806]: E1204 03:55:21.423403 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:21 crc kubenswrapper[4806]: E1204 03:55:21.423444 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:21 crc kubenswrapper[4806]: E1204 03:55:21.423349 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:21 crc kubenswrapper[4806]: E1204 03:55:21.423493 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.466569 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs\") pod \"network-metrics-daemon-b9khd\" (UID: \"41fdb600-3868-48ab-8396-17c2382b6168\") " pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:21 crc kubenswrapper[4806]: E1204 03:55:21.466755 4806 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 03:55:21 crc kubenswrapper[4806]: E1204 03:55:21.466813 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs podName:41fdb600-3868-48ab-8396-17c2382b6168 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:23.466796156 +0000 UTC m=+38.325309094 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs") pod "network-metrics-daemon-b9khd" (UID: "41fdb600-3868-48ab-8396-17c2382b6168") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.482347 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.482378 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.482387 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.482402 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.482413 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:21Z","lastTransitionTime":"2025-12-04T03:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.584896 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.584992 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.585006 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.585022 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.585034 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:21Z","lastTransitionTime":"2025-12-04T03:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.687279 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.687310 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.687318 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.687330 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.687339 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:21Z","lastTransitionTime":"2025-12-04T03:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.790252 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.790292 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.790303 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.790318 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.790330 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:21Z","lastTransitionTime":"2025-12-04T03:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.892897 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.892968 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.892977 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.892991 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.893001 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:21Z","lastTransitionTime":"2025-12-04T03:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.995190 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.995288 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.995301 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.995321 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:21 crc kubenswrapper[4806]: I1204 03:55:21.995337 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:21Z","lastTransitionTime":"2025-12-04T03:55:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.098423 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.098493 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.098510 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.098533 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.098550 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:22Z","lastTransitionTime":"2025-12-04T03:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.200635 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.200691 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.200706 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.200727 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.200743 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:22Z","lastTransitionTime":"2025-12-04T03:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.303804 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.303860 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.303877 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.303899 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.303914 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:22Z","lastTransitionTime":"2025-12-04T03:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.406108 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.406157 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.406168 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.406182 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.406193 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:22Z","lastTransitionTime":"2025-12-04T03:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.508750 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.508816 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.508831 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.508854 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.509245 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:22Z","lastTransitionTime":"2025-12-04T03:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.611961 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.612007 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.612017 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.612034 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.612045 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:22Z","lastTransitionTime":"2025-12-04T03:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.715332 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.715394 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.715406 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.715425 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.715439 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:22Z","lastTransitionTime":"2025-12-04T03:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.818818 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.818885 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.818903 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.818963 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.818984 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:22Z","lastTransitionTime":"2025-12-04T03:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.921880 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.921988 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.922013 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.922078 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:22 crc kubenswrapper[4806]: I1204 03:55:22.922096 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:22Z","lastTransitionTime":"2025-12-04T03:55:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.024144 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.024195 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.024206 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.024221 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.024242 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:23Z","lastTransitionTime":"2025-12-04T03:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.126766 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.126815 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.126825 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.126849 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.126869 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:23Z","lastTransitionTime":"2025-12-04T03:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.230180 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.230247 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.230263 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.230285 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.230300 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:23Z","lastTransitionTime":"2025-12-04T03:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.333353 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.333398 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.333409 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.333424 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.333436 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:23Z","lastTransitionTime":"2025-12-04T03:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.423162 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.423230 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.423188 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:23 crc kubenswrapper[4806]: E1204 03:55:23.423329 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.423259 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:23 crc kubenswrapper[4806]: E1204 03:55:23.423590 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:23 crc kubenswrapper[4806]: E1204 03:55:23.423720 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:23 crc kubenswrapper[4806]: E1204 03:55:23.423907 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.435282 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.435361 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.435403 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.435427 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.435450 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:23Z","lastTransitionTime":"2025-12-04T03:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.490368 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs\") pod \"network-metrics-daemon-b9khd\" (UID: \"41fdb600-3868-48ab-8396-17c2382b6168\") " pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:23 crc kubenswrapper[4806]: E1204 03:55:23.490492 4806 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 03:55:23 crc kubenswrapper[4806]: E1204 03:55:23.490578 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs podName:41fdb600-3868-48ab-8396-17c2382b6168 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:27.49055977 +0000 UTC m=+42.349072708 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs") pod "network-metrics-daemon-b9khd" (UID: "41fdb600-3868-48ab-8396-17c2382b6168") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.537322 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.537376 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.537387 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.537439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.537449 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:23Z","lastTransitionTime":"2025-12-04T03:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.639467 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.639510 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.639524 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.639543 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.639553 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:23Z","lastTransitionTime":"2025-12-04T03:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.742105 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.742160 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.742169 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.742203 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.742212 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:23Z","lastTransitionTime":"2025-12-04T03:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.844440 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.844488 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.844524 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.844561 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.844572 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:23Z","lastTransitionTime":"2025-12-04T03:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.947712 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.947803 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.947832 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.947905 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:23 crc kubenswrapper[4806]: I1204 03:55:23.947972 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:23Z","lastTransitionTime":"2025-12-04T03:55:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.050904 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.050973 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.050985 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.051000 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.051038 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:24Z","lastTransitionTime":"2025-12-04T03:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.153657 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.153707 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.153721 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.153740 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.153755 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:24Z","lastTransitionTime":"2025-12-04T03:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.256094 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.256168 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.256187 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.256215 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.256234 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:24Z","lastTransitionTime":"2025-12-04T03:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.358650 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.358692 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.358702 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.358717 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.358728 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:24Z","lastTransitionTime":"2025-12-04T03:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.461868 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.461902 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.461910 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.461943 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.461953 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:24Z","lastTransitionTime":"2025-12-04T03:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.565427 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.565474 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.565482 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.565497 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.565506 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:24Z","lastTransitionTime":"2025-12-04T03:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.667818 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.667859 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.667871 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.667885 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.667895 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:24Z","lastTransitionTime":"2025-12-04T03:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.771079 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.771125 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.771134 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.771148 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.771158 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:24Z","lastTransitionTime":"2025-12-04T03:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.877037 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.877092 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.877113 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.877131 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.877142 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:24Z","lastTransitionTime":"2025-12-04T03:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.980156 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.980218 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.980238 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.980262 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:24 crc kubenswrapper[4806]: I1204 03:55:24.980280 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:24Z","lastTransitionTime":"2025-12-04T03:55:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.082997 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.083040 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.083050 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.083067 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.083078 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:25Z","lastTransitionTime":"2025-12-04T03:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.185300 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.185387 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.185400 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.185417 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.185461 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:25Z","lastTransitionTime":"2025-12-04T03:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.287633 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.288006 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.288091 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.288203 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.288295 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:25Z","lastTransitionTime":"2025-12-04T03:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.391228 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.391299 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.391320 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.391348 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.391368 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:25Z","lastTransitionTime":"2025-12-04T03:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.422683 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:25 crc kubenswrapper[4806]: E1204 03:55:25.423207 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.423264 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.423417 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:25 crc kubenswrapper[4806]: E1204 03:55:25.423482 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:25 crc kubenswrapper[4806]: E1204 03:55:25.423536 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.423858 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:25 crc kubenswrapper[4806]: E1204 03:55:25.424228 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.445478 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:25Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.463203 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:25Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.477059 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:25Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.493433 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.493695 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.493839 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.494011 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.494095 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:25Z","lastTransitionTime":"2025-12-04T03:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.498328 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:25Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.519452 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:25Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.530699 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:25Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.555747 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\":55:16.345209 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 03:55:16.345217 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 03:55:16.345224 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 03:55:16.345231 6103 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 03:55:16.345302 6103 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345448 6103 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 03:55:16.345527 6103 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345754 6103 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345963 6103 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346267 6103 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346350 6103 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:25Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.568821 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b9khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41fdb600-3868-48ab-8396-17c2382b6168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b9khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:25Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.581792 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:25Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.596845 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.596891 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.596908 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.596950 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.596966 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:25Z","lastTransitionTime":"2025-12-04T03:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.601223 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:25Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.619484 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:25Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.635891 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.636018 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:25Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.637434 4806 scope.go:117] "RemoveContainer" containerID="1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297" Dec 04 03:55:25 crc kubenswrapper[4806]: E1204 03:55:25.637702 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.653575 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:25Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.671181 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:25Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.689065 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:25Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.699084 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.699120 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.699131 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.699147 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.699158 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:25Z","lastTransitionTime":"2025-12-04T03:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.702133 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:25Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.802534 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.802601 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.802619 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.802644 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.802666 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:25Z","lastTransitionTime":"2025-12-04T03:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.905888 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.906006 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.906030 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.906058 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:25 crc kubenswrapper[4806]: I1204 03:55:25.906080 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:25Z","lastTransitionTime":"2025-12-04T03:55:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.009073 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.009199 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.009212 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.009228 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.009237 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:26Z","lastTransitionTime":"2025-12-04T03:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.111858 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.111913 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.111938 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.111953 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.111964 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:26Z","lastTransitionTime":"2025-12-04T03:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.214246 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.214279 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.214287 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.214300 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.214309 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:26Z","lastTransitionTime":"2025-12-04T03:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.317762 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.317869 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.317907 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.317987 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.318029 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:26Z","lastTransitionTime":"2025-12-04T03:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.420443 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.420535 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.420563 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.420593 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.420634 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:26Z","lastTransitionTime":"2025-12-04T03:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.524188 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.524256 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.524274 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.524295 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.524312 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:26Z","lastTransitionTime":"2025-12-04T03:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.626359 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.626405 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.626419 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.626438 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.626453 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:26Z","lastTransitionTime":"2025-12-04T03:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.729112 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.729156 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.729176 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.729198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.729214 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:26Z","lastTransitionTime":"2025-12-04T03:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.832839 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.832890 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.832904 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.832947 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.832961 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:26Z","lastTransitionTime":"2025-12-04T03:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.935540 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.935581 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.935596 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.935614 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:26 crc kubenswrapper[4806]: I1204 03:55:26.935627 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:26Z","lastTransitionTime":"2025-12-04T03:55:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.039326 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.039443 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.039468 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.039499 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.039521 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:27Z","lastTransitionTime":"2025-12-04T03:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.102629 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.102675 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.102686 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.102713 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.102726 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:27Z","lastTransitionTime":"2025-12-04T03:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:27 crc kubenswrapper[4806]: E1204 03:55:27.117748 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:27Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.127771 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.127843 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.127867 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.127892 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.127905 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:27Z","lastTransitionTime":"2025-12-04T03:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:27 crc kubenswrapper[4806]: E1204 03:55:27.141888 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:27Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.146241 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.146282 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.146291 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.146306 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.146316 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:27Z","lastTransitionTime":"2025-12-04T03:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:27 crc kubenswrapper[4806]: E1204 03:55:27.158917 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:27Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.163072 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.163161 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.163174 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.163193 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.163205 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:27Z","lastTransitionTime":"2025-12-04T03:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:27 crc kubenswrapper[4806]: E1204 03:55:27.176479 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:27Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.180419 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.180458 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.180469 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.180485 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.180496 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:27Z","lastTransitionTime":"2025-12-04T03:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:27 crc kubenswrapper[4806]: E1204 03:55:27.194638 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:27Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:27 crc kubenswrapper[4806]: E1204 03:55:27.194759 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.196300 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.196332 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.196347 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.196362 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.196373 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:27Z","lastTransitionTime":"2025-12-04T03:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.299229 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.299269 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.299279 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.299294 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.299332 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:27Z","lastTransitionTime":"2025-12-04T03:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.401674 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.401727 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.401737 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.401752 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.401762 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:27Z","lastTransitionTime":"2025-12-04T03:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.423146 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:27 crc kubenswrapper[4806]: E1204 03:55:27.423314 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.423351 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.423379 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.423147 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:27 crc kubenswrapper[4806]: E1204 03:55:27.423431 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:27 crc kubenswrapper[4806]: E1204 03:55:27.423555 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:27 crc kubenswrapper[4806]: E1204 03:55:27.423677 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.504489 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.504520 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.504530 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.504546 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.504556 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:27Z","lastTransitionTime":"2025-12-04T03:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.534768 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs\") pod \"network-metrics-daemon-b9khd\" (UID: \"41fdb600-3868-48ab-8396-17c2382b6168\") " pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:27 crc kubenswrapper[4806]: E1204 03:55:27.534895 4806 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 03:55:27 crc kubenswrapper[4806]: E1204 03:55:27.535005 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs podName:41fdb600-3868-48ab-8396-17c2382b6168 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:35.534983173 +0000 UTC m=+50.393496111 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs") pod "network-metrics-daemon-b9khd" (UID: "41fdb600-3868-48ab-8396-17c2382b6168") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.606866 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.606900 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.606909 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.606939 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.606950 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:27Z","lastTransitionTime":"2025-12-04T03:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.709519 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.709591 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.709610 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.709638 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.709657 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:27Z","lastTransitionTime":"2025-12-04T03:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.812137 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.812219 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.812245 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.812278 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.812302 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:27Z","lastTransitionTime":"2025-12-04T03:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.916313 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.916366 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.916382 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.916404 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:27 crc kubenswrapper[4806]: I1204 03:55:27.916420 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:27Z","lastTransitionTime":"2025-12-04T03:55:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.019374 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.019441 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.019466 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.019506 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.019532 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:28Z","lastTransitionTime":"2025-12-04T03:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.122994 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.123060 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.123071 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.123088 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.123098 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:28Z","lastTransitionTime":"2025-12-04T03:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.225621 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.225658 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.225669 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.225683 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.225695 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:28Z","lastTransitionTime":"2025-12-04T03:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.327880 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.327987 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.328006 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.328030 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.328048 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:28Z","lastTransitionTime":"2025-12-04T03:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.430481 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.430794 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.430890 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.431041 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.431167 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:28Z","lastTransitionTime":"2025-12-04T03:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.534315 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.534387 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.534412 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.534435 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.534452 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:28Z","lastTransitionTime":"2025-12-04T03:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.637458 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.637899 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.638126 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.638361 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.638570 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:28Z","lastTransitionTime":"2025-12-04T03:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.742114 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.742990 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.743308 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.743523 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.743741 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:28Z","lastTransitionTime":"2025-12-04T03:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.846582 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.847078 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.847326 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.847557 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.847766 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:28Z","lastTransitionTime":"2025-12-04T03:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.950222 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.950555 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.950648 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.950740 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:28 crc kubenswrapper[4806]: I1204 03:55:28.950818 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:28Z","lastTransitionTime":"2025-12-04T03:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.053475 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.053510 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.053522 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.053538 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.053549 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:29Z","lastTransitionTime":"2025-12-04T03:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.156669 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.156750 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.156774 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.156803 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.156825 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:29Z","lastTransitionTime":"2025-12-04T03:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.259451 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.259490 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.259500 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.259515 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.259526 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:29Z","lastTransitionTime":"2025-12-04T03:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.362226 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.362313 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.362334 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.362356 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.362371 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:29Z","lastTransitionTime":"2025-12-04T03:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.423336 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.423497 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.423540 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.423770 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:29 crc kubenswrapper[4806]: E1204 03:55:29.423843 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:29 crc kubenswrapper[4806]: E1204 03:55:29.423946 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:29 crc kubenswrapper[4806]: E1204 03:55:29.424107 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:29 crc kubenswrapper[4806]: E1204 03:55:29.424126 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.464866 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.465170 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.465249 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.465332 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.465407 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:29Z","lastTransitionTime":"2025-12-04T03:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.568117 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.568195 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.568216 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.568244 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.568266 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:29Z","lastTransitionTime":"2025-12-04T03:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.670401 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.670446 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.670456 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.670473 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.670485 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:29Z","lastTransitionTime":"2025-12-04T03:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.773062 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.773100 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.773112 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.773127 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.773139 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:29Z","lastTransitionTime":"2025-12-04T03:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.875126 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.875699 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.875779 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.875880 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.875967 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:29Z","lastTransitionTime":"2025-12-04T03:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.978471 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.978518 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.978527 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.978540 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:29 crc kubenswrapper[4806]: I1204 03:55:29.978549 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:29Z","lastTransitionTime":"2025-12-04T03:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.080624 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.080672 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.080681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.080695 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.080705 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:30Z","lastTransitionTime":"2025-12-04T03:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.183199 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.183685 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.183893 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.184105 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.184287 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:30Z","lastTransitionTime":"2025-12-04T03:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.287322 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.287633 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.287724 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.287807 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.287868 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:30Z","lastTransitionTime":"2025-12-04T03:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.390178 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.390491 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.390581 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.390661 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.390825 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:30Z","lastTransitionTime":"2025-12-04T03:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.493857 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.493941 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.493966 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.493989 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.494002 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:30Z","lastTransitionTime":"2025-12-04T03:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.596535 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.596577 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.596586 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.596600 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.596612 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:30Z","lastTransitionTime":"2025-12-04T03:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.698468 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.698886 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.698984 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.699070 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.699148 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:30Z","lastTransitionTime":"2025-12-04T03:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.801630 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.801682 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.801691 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.801705 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.801716 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:30Z","lastTransitionTime":"2025-12-04T03:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.904139 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.904421 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.904480 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.904581 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:30 crc kubenswrapper[4806]: I1204 03:55:30.904645 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:30Z","lastTransitionTime":"2025-12-04T03:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.007899 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.007978 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.007995 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.008021 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.008062 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:31Z","lastTransitionTime":"2025-12-04T03:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.110991 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.111045 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.111058 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.111084 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.111100 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:31Z","lastTransitionTime":"2025-12-04T03:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.214396 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.214452 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.214465 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.214484 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.214498 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:31Z","lastTransitionTime":"2025-12-04T03:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.317852 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.318250 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.318342 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.318420 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.318610 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:31Z","lastTransitionTime":"2025-12-04T03:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.421981 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.422434 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.422545 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.422671 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.422806 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.422769 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.422829 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.423056 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.422851 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:31Z","lastTransitionTime":"2025-12-04T03:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:31 crc kubenswrapper[4806]: E1204 03:55:31.423342 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:31 crc kubenswrapper[4806]: E1204 03:55:31.423232 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:31 crc kubenswrapper[4806]: E1204 03:55:31.423100 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:31 crc kubenswrapper[4806]: E1204 03:55:31.423521 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.526745 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.526804 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.526818 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.526836 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.526849 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:31Z","lastTransitionTime":"2025-12-04T03:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.630700 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.631133 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.631224 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.631347 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.631449 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:31Z","lastTransitionTime":"2025-12-04T03:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.736271 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.736344 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.736369 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.736403 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.736563 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:31Z","lastTransitionTime":"2025-12-04T03:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.839843 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.839889 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.839897 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.839912 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.839935 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:31Z","lastTransitionTime":"2025-12-04T03:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.942420 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.942461 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.942472 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.942487 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:31 crc kubenswrapper[4806]: I1204 03:55:31.942495 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:31Z","lastTransitionTime":"2025-12-04T03:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.046001 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.046047 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.046060 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.046073 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.046104 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:32Z","lastTransitionTime":"2025-12-04T03:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.148246 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.148287 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.148298 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.148313 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.148324 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:32Z","lastTransitionTime":"2025-12-04T03:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.250916 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.250973 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.250984 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.250998 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.251008 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:32Z","lastTransitionTime":"2025-12-04T03:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.353092 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.353151 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.353164 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.353181 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.353194 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:32Z","lastTransitionTime":"2025-12-04T03:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.457508 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.457912 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.458025 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.458140 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.458365 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:32Z","lastTransitionTime":"2025-12-04T03:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.561018 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.561270 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.561347 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.561578 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.561682 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:32Z","lastTransitionTime":"2025-12-04T03:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.664060 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.664101 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.664109 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.664123 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.664133 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:32Z","lastTransitionTime":"2025-12-04T03:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.767973 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.768042 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.768058 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.768083 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.768099 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:32Z","lastTransitionTime":"2025-12-04T03:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.870818 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.870883 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.870905 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.870966 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.870992 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:32Z","lastTransitionTime":"2025-12-04T03:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.973642 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.973755 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.973780 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.973845 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:32 crc kubenswrapper[4806]: I1204 03:55:32.973871 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:32Z","lastTransitionTime":"2025-12-04T03:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.076363 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.076432 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.076441 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.076453 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.076461 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:33Z","lastTransitionTime":"2025-12-04T03:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.179624 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.179687 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.179715 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.179742 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.179768 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:33Z","lastTransitionTime":"2025-12-04T03:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.282858 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.282897 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.282907 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.282941 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.282954 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:33Z","lastTransitionTime":"2025-12-04T03:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.386050 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.386117 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.386128 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.386144 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.386155 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:33Z","lastTransitionTime":"2025-12-04T03:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.423766 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.423850 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:33 crc kubenswrapper[4806]: E1204 03:55:33.423967 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.423993 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:33 crc kubenswrapper[4806]: E1204 03:55:33.424177 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:33 crc kubenswrapper[4806]: E1204 03:55:33.424273 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.424464 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:33 crc kubenswrapper[4806]: E1204 03:55:33.424746 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.489703 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.489744 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.489758 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.489775 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.489788 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:33Z","lastTransitionTime":"2025-12-04T03:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.593160 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.593232 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.593271 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.593305 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.593329 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:33Z","lastTransitionTime":"2025-12-04T03:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.698886 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.699326 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.699479 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.699690 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.699845 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:33Z","lastTransitionTime":"2025-12-04T03:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.803256 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.803308 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.803324 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.803347 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.803365 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:33Z","lastTransitionTime":"2025-12-04T03:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.907246 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.907310 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.907338 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.907367 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:33 crc kubenswrapper[4806]: I1204 03:55:33.907389 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:33Z","lastTransitionTime":"2025-12-04T03:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.011343 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.011762 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.012012 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.012171 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.012269 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:34Z","lastTransitionTime":"2025-12-04T03:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.116111 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.116173 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.116186 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.116205 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.116218 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:34Z","lastTransitionTime":"2025-12-04T03:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.219741 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.220261 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.220488 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.220738 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.220999 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:34Z","lastTransitionTime":"2025-12-04T03:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.323717 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.324119 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.324308 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.324448 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.324566 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:34Z","lastTransitionTime":"2025-12-04T03:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.427439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.427492 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.427503 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.427521 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.427533 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:34Z","lastTransitionTime":"2025-12-04T03:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.530349 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.530425 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.530447 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.530476 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.530498 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:34Z","lastTransitionTime":"2025-12-04T03:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.634179 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.634250 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.634267 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.634292 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.634309 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:34Z","lastTransitionTime":"2025-12-04T03:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.736234 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.736276 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.736305 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.736323 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.736332 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:34Z","lastTransitionTime":"2025-12-04T03:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.839447 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.839481 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.839489 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.839503 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.839513 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:34Z","lastTransitionTime":"2025-12-04T03:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.942887 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.942967 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.942983 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.943004 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:34 crc kubenswrapper[4806]: I1204 03:55:34.943021 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:34Z","lastTransitionTime":"2025-12-04T03:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.045547 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.045585 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.045594 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.045611 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.045623 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:35Z","lastTransitionTime":"2025-12-04T03:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.147751 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.147819 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.147841 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.147870 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.147890 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:35Z","lastTransitionTime":"2025-12-04T03:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.216609 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.216692 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.216845 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.216903 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.216846 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.217009 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.216970 4806 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.217037 4806 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.217219 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 03:56:07.217181836 +0000 UTC m=+82.075694804 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.217338 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 03:56:07.21730844 +0000 UTC m=+82.075821398 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.249895 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.250046 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.250063 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.250079 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.250365 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:35Z","lastTransitionTime":"2025-12-04T03:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.317472 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.317605 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.317650 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.317706 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:56:07.317672389 +0000 UTC m=+82.176185317 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.317748 4806 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.317793 4806 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.317817 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 03:56:07.317800903 +0000 UTC m=+82.176313881 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.317854 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 03:56:07.317842124 +0000 UTC m=+82.176355152 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.352738 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.352794 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.352805 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.352826 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.352840 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:35Z","lastTransitionTime":"2025-12-04T03:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.422876 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.422905 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.423102 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.423223 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.423375 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.423533 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.423674 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.423769 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.437383 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:35Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.446366 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:35Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.455573 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.455609 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.455618 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.455632 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.455642 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:35Z","lastTransitionTime":"2025-12-04T03:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.462118 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:35Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.475764 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:35Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.491191 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:35Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.511885 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:35Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.530271 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:35Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.543625 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:35Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.555910 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:35Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.562750 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.562798 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.562810 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.562830 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.562847 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:35Z","lastTransitionTime":"2025-12-04T03:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.575280 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:35Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.585346 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:35Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.594744 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b9khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41fdb600-3868-48ab-8396-17c2382b6168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b9khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:35Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.605885 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:35Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.616973 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:35Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.621378 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs\") pod \"network-metrics-daemon-b9khd\" (UID: \"41fdb600-3868-48ab-8396-17c2382b6168\") " pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.621646 4806 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 03:55:35 crc kubenswrapper[4806]: E1204 03:55:35.621750 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs podName:41fdb600-3868-48ab-8396-17c2382b6168 nodeName:}" failed. No retries permitted until 2025-12-04 03:55:51.62172953 +0000 UTC m=+66.480242468 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs") pod "network-metrics-daemon-b9khd" (UID: "41fdb600-3868-48ab-8396-17c2382b6168") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.627081 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:35Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.647393 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\":55:16.345209 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 03:55:16.345217 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 03:55:16.345224 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 03:55:16.345231 6103 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 03:55:16.345302 6103 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345448 6103 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 03:55:16.345527 6103 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345754 6103 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345963 6103 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346267 6103 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346350 6103 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:35Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.668060 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.668108 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.668116 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.668129 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.668138 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:35Z","lastTransitionTime":"2025-12-04T03:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.770957 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.771253 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.771329 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.771398 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.771472 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:35Z","lastTransitionTime":"2025-12-04T03:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.873635 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.873668 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.873677 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.873690 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.873699 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:35Z","lastTransitionTime":"2025-12-04T03:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.977243 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.977324 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.977350 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.977383 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:35 crc kubenswrapper[4806]: I1204 03:55:35.977405 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:35Z","lastTransitionTime":"2025-12-04T03:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.080284 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.080335 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.080350 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.080372 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.080387 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:36Z","lastTransitionTime":"2025-12-04T03:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.184392 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.184428 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.184439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.184453 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.184464 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:36Z","lastTransitionTime":"2025-12-04T03:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.287474 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.287515 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.287529 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.287545 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.287556 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:36Z","lastTransitionTime":"2025-12-04T03:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.391013 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.391079 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.391092 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.391113 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.391125 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:36Z","lastTransitionTime":"2025-12-04T03:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.494333 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.494406 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.494419 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.494441 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.494461 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:36Z","lastTransitionTime":"2025-12-04T03:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.597598 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.597645 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.597656 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.597671 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.597683 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:36Z","lastTransitionTime":"2025-12-04T03:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.699843 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.699898 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.699952 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.699978 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.699995 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:36Z","lastTransitionTime":"2025-12-04T03:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.802972 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.803015 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.803024 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.803040 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.803049 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:36Z","lastTransitionTime":"2025-12-04T03:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.905812 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.905873 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.905891 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.905910 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:36 crc kubenswrapper[4806]: I1204 03:55:36.905939 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:36Z","lastTransitionTime":"2025-12-04T03:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.008388 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.008450 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.008464 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.008483 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.008508 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:37Z","lastTransitionTime":"2025-12-04T03:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.110959 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.111004 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.111015 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.111029 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.111039 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:37Z","lastTransitionTime":"2025-12-04T03:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.213410 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.213448 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.213459 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.213473 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.213482 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:37Z","lastTransitionTime":"2025-12-04T03:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.316712 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.316764 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.316781 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.316801 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.316817 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:37Z","lastTransitionTime":"2025-12-04T03:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.419021 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.419501 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.419659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.419796 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.419917 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:37Z","lastTransitionTime":"2025-12-04T03:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.423176 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.423189 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.423204 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.423795 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:37 crc kubenswrapper[4806]: E1204 03:55:37.423960 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.423871 4806 scope.go:117] "RemoveContainer" containerID="1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297" Dec 04 03:55:37 crc kubenswrapper[4806]: E1204 03:55:37.424451 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:37 crc kubenswrapper[4806]: E1204 03:55:37.424531 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:37 crc kubenswrapper[4806]: E1204 03:55:37.424976 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.462256 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.462294 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.462307 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.462325 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.462349 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:37Z","lastTransitionTime":"2025-12-04T03:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:37 crc kubenswrapper[4806]: E1204 03:55:37.477558 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.481494 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.481971 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.482137 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.482275 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.482861 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:37Z","lastTransitionTime":"2025-12-04T03:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:37 crc kubenswrapper[4806]: E1204 03:55:37.496104 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.500692 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.500962 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.501108 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.501237 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.501358 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:37Z","lastTransitionTime":"2025-12-04T03:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:37 crc kubenswrapper[4806]: E1204 03:55:37.515884 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.520273 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.520320 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.520337 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.520358 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.520373 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:37Z","lastTransitionTime":"2025-12-04T03:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:37 crc kubenswrapper[4806]: E1204 03:55:37.533350 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.537419 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.537456 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.537467 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.537484 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.537495 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:37Z","lastTransitionTime":"2025-12-04T03:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:37 crc kubenswrapper[4806]: E1204 03:55:37.551293 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: E1204 03:55:37.551457 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.553080 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.553139 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.553153 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.553171 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.553183 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:37Z","lastTransitionTime":"2025-12-04T03:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.656957 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.657212 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.657224 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.657238 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.657248 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:37Z","lastTransitionTime":"2025-12-04T03:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.710477 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/1.log" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.713521 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerStarted","Data":"d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4"} Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.715412 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.727423 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.738706 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.750307 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.765439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.765475 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.765484 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.765498 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.765513 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:37Z","lastTransitionTime":"2025-12-04T03:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.770397 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.789950 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.808743 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.833648 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.847988 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.866420 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\":55:16.345209 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 03:55:16.345217 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 03:55:16.345224 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 03:55:16.345231 6103 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 03:55:16.345302 6103 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345448 6103 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 03:55:16.345527 6103 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345754 6103 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345963 6103 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346267 6103 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346350 6103 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.867238 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.867276 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.867284 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.867299 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.867311 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:37Z","lastTransitionTime":"2025-12-04T03:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.875811 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b9khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41fdb600-3868-48ab-8396-17c2382b6168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b9khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.887568 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.899086 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.912458 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.923644 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.936947 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.951652 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:37Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.969385 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.969431 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.969442 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.969459 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:37 crc kubenswrapper[4806]: I1204 03:55:37.969470 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:37Z","lastTransitionTime":"2025-12-04T03:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.071142 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.071178 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.071189 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.071204 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.071214 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:38Z","lastTransitionTime":"2025-12-04T03:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.173964 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.174010 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.174020 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.174035 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.174045 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:38Z","lastTransitionTime":"2025-12-04T03:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.277439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.277501 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.277529 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.277544 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.277553 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:38Z","lastTransitionTime":"2025-12-04T03:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.379938 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.379981 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.379991 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.380003 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.380012 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:38Z","lastTransitionTime":"2025-12-04T03:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.482517 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.482578 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.482590 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.482605 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.482616 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:38Z","lastTransitionTime":"2025-12-04T03:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.586439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.586511 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.586533 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.586561 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.586584 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:38Z","lastTransitionTime":"2025-12-04T03:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.688862 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.688914 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.688959 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.688977 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.688989 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:38Z","lastTransitionTime":"2025-12-04T03:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.718647 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/2.log" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.719352 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/1.log" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.722589 4806 generic.go:334] "Generic (PLEG): container finished" podID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerID="d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4" exitCode=1 Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.722642 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerDied","Data":"d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4"} Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.722692 4806 scope.go:117] "RemoveContainer" containerID="1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.725113 4806 scope.go:117] "RemoveContainer" containerID="d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4" Dec 04 03:55:38 crc kubenswrapper[4806]: E1204 03:55:38.725514 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.745444 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:38Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.760539 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:38Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.775918 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:38Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.792205 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.792240 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.792249 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.792262 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.792272 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:38Z","lastTransitionTime":"2025-12-04T03:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.792687 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:38Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.807654 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:38Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.819242 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:38Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.837760 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\":55:16.345209 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 03:55:16.345217 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 03:55:16.345224 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 03:55:16.345231 6103 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 03:55:16.345302 6103 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345448 6103 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 03:55:16.345527 6103 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345754 6103 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345963 6103 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346267 6103 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346350 6103 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:38Z\\\",\\\"message\\\":\\\"1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 03:55:38.189705 6360 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1204 03:55:38.189692 6360 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 03:55:38.189728 6360 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 03:55:38.189313 6360 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1204 03:55:38.189731 6360 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Tim\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:38Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.848270 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b9khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41fdb600-3868-48ab-8396-17c2382b6168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b9khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:38Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.863009 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:38Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.876981 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:38Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.891145 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:38Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.895697 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.896082 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.896182 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.896299 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.896436 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:38Z","lastTransitionTime":"2025-12-04T03:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.910290 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:38Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.923188 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:38Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.934416 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:38Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.946422 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:38Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.957965 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:38Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.999509 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.999538 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.999546 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.999558 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:38 crc kubenswrapper[4806]: I1204 03:55:38.999566 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:38Z","lastTransitionTime":"2025-12-04T03:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.102975 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.103035 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.103056 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.103082 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.103099 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:39Z","lastTransitionTime":"2025-12-04T03:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.205462 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.205499 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.205509 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.205525 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.205536 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:39Z","lastTransitionTime":"2025-12-04T03:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.239515 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.252406 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.254621 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.267889 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.279432 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.293882 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.304062 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.307659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.307697 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.307730 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.307745 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.307758 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:39Z","lastTransitionTime":"2025-12-04T03:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.323321 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e1f07e6efe766da2fca234bd36da55c6ef31ca234d89e9b35bd59d29f9e1297\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:16Z\\\",\\\"message\\\":\\\":55:16.345209 6103 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 03:55:16.345217 6103 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1204 03:55:16.345224 6103 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1204 03:55:16.345231 6103 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 03:55:16.345302 6103 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345448 6103 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 03:55:16.345527 6103 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345754 6103 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.345963 6103 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346267 6103 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 03:55:16.346350 6103 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:38Z\\\",\\\"message\\\":\\\"1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 03:55:38.189705 6360 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1204 03:55:38.189692 6360 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 03:55:38.189728 6360 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 03:55:38.189313 6360 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1204 03:55:38.189731 6360 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Tim\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.336598 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b9khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41fdb600-3868-48ab-8396-17c2382b6168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b9khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.347841 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.367071 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.383345 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.395896 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.411021 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.411140 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.411162 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.411300 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.411338 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:39Z","lastTransitionTime":"2025-12-04T03:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.414171 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.423079 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.423118 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.423249 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:39 crc kubenswrapper[4806]: E1204 03:55:39.423339 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.423404 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:39 crc kubenswrapper[4806]: E1204 03:55:39.423576 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:39 crc kubenswrapper[4806]: E1204 03:55:39.423627 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:39 crc kubenswrapper[4806]: E1204 03:55:39.423689 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.429068 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.442197 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.454833 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.466494 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.515725 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.515776 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.515788 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.515805 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.515818 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:39Z","lastTransitionTime":"2025-12-04T03:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.618687 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.618798 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.618824 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.618840 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.618848 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:39Z","lastTransitionTime":"2025-12-04T03:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.722503 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.722541 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.722550 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.722563 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.722572 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:39Z","lastTransitionTime":"2025-12-04T03:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.728062 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/2.log" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.732399 4806 scope.go:117] "RemoveContainer" containerID="d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4" Dec 04 03:55:39 crc kubenswrapper[4806]: E1204 03:55:39.732547 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.746198 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.758731 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.769850 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.781795 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.792643 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.803095 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.820254 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:38Z\\\",\\\"message\\\":\\\"1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 03:55:38.189705 6360 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1204 03:55:38.189692 6360 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 03:55:38.189728 6360 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 03:55:38.189313 6360 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1204 03:55:38.189731 6360 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Tim\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.824625 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.824655 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.824665 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.824680 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.824691 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:39Z","lastTransitionTime":"2025-12-04T03:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.830346 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b9khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41fdb600-3868-48ab-8396-17c2382b6168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b9khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.843312 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.853902 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"186beef2-bf40-41df-ad1d-870b99e5b6f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59760c3aea1ddf4bdfeda36dd7e6e20bb56985e39f27864475a064d0061325de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd1df0baab4bc2ca707440eaa20c66a5b003a783d48ac8ed0c66d185a6e4d116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4e4be31ada4276e640bcf04ab82968c378cfaa920bc82ce54d925d457c51585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.875732 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.896306 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.922321 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.926874 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.926916 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.926940 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.926953 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.926962 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:39Z","lastTransitionTime":"2025-12-04T03:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.939079 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.949734 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.960656 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:39 crc kubenswrapper[4806]: I1204 03:55:39.973861 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:39Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.030180 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.030244 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.030257 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.030589 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.030791 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:40Z","lastTransitionTime":"2025-12-04T03:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.134447 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.134539 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.134558 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.134620 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.134640 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:40Z","lastTransitionTime":"2025-12-04T03:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.237013 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.237072 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.237092 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.237114 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.237131 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:40Z","lastTransitionTime":"2025-12-04T03:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.339784 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.339844 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.339857 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.339876 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.339887 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:40Z","lastTransitionTime":"2025-12-04T03:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.443298 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.443364 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.443375 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.443392 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.443403 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:40Z","lastTransitionTime":"2025-12-04T03:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.546215 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.546262 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.546277 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.546293 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.546304 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:40Z","lastTransitionTime":"2025-12-04T03:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.648688 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.648738 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.648750 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.648768 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.648779 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:40Z","lastTransitionTime":"2025-12-04T03:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.750777 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.750819 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.750830 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.750846 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.750857 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:40Z","lastTransitionTime":"2025-12-04T03:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.853703 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.853776 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.853801 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.853833 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.853856 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:40Z","lastTransitionTime":"2025-12-04T03:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.956893 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.957012 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.957081 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.957115 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:40 crc kubenswrapper[4806]: I1204 03:55:40.957142 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:40Z","lastTransitionTime":"2025-12-04T03:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.060023 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.060078 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.060091 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.060111 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.060122 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:41Z","lastTransitionTime":"2025-12-04T03:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.163260 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.163334 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.163358 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.163392 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.163414 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:41Z","lastTransitionTime":"2025-12-04T03:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.267089 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.267140 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.267152 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.267171 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.267185 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:41Z","lastTransitionTime":"2025-12-04T03:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.370017 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.370114 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.370127 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.370152 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.370167 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:41Z","lastTransitionTime":"2025-12-04T03:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.423796 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.423903 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.423795 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:41 crc kubenswrapper[4806]: E1204 03:55:41.424073 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:41 crc kubenswrapper[4806]: E1204 03:55:41.424302 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:41 crc kubenswrapper[4806]: E1204 03:55:41.424396 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.424539 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:41 crc kubenswrapper[4806]: E1204 03:55:41.424664 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.472814 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.472859 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.472875 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.472896 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.472911 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:41Z","lastTransitionTime":"2025-12-04T03:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.576237 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.576283 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.576294 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.576308 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.576317 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:41Z","lastTransitionTime":"2025-12-04T03:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.679215 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.679482 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.679508 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.679548 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.679570 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:41Z","lastTransitionTime":"2025-12-04T03:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.787274 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.787321 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.787381 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.787397 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.787408 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:41Z","lastTransitionTime":"2025-12-04T03:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.891799 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.891849 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.891860 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.891885 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.891900 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:41Z","lastTransitionTime":"2025-12-04T03:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.994768 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.994837 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.994849 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.994871 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:41 crc kubenswrapper[4806]: I1204 03:55:41.994885 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:41Z","lastTransitionTime":"2025-12-04T03:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.097209 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.097373 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.097389 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.097407 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.097419 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:42Z","lastTransitionTime":"2025-12-04T03:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.199853 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.199889 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.199897 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.199912 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.199945 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:42Z","lastTransitionTime":"2025-12-04T03:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.302351 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.302399 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.302411 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.302426 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.302437 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:42Z","lastTransitionTime":"2025-12-04T03:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.405405 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.405467 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.405484 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.405503 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.405515 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:42Z","lastTransitionTime":"2025-12-04T03:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.508047 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.508115 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.508129 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.508146 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.508157 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:42Z","lastTransitionTime":"2025-12-04T03:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.611483 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.611521 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.611529 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.611541 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.611549 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:42Z","lastTransitionTime":"2025-12-04T03:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.715248 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.715325 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.715346 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.715375 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.715398 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:42Z","lastTransitionTime":"2025-12-04T03:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.818079 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.818116 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.818124 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.818136 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.818145 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:42Z","lastTransitionTime":"2025-12-04T03:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.920322 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.920365 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.920373 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.920386 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:42 crc kubenswrapper[4806]: I1204 03:55:42.920395 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:42Z","lastTransitionTime":"2025-12-04T03:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.022805 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.022852 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.022862 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.022880 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.022893 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:43Z","lastTransitionTime":"2025-12-04T03:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.125442 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.125478 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.125486 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.125498 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.125507 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:43Z","lastTransitionTime":"2025-12-04T03:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.227884 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.227953 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.227968 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.227989 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.228004 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:43Z","lastTransitionTime":"2025-12-04T03:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.330877 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.330918 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.330949 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.330965 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.330974 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:43Z","lastTransitionTime":"2025-12-04T03:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.423090 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.423203 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.423368 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:43 crc kubenswrapper[4806]: E1204 03:55:43.423369 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.423388 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:43 crc kubenswrapper[4806]: E1204 03:55:43.423464 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:43 crc kubenswrapper[4806]: E1204 03:55:43.423480 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:43 crc kubenswrapper[4806]: E1204 03:55:43.423521 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.432598 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.432675 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.432691 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.432708 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.432718 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:43Z","lastTransitionTime":"2025-12-04T03:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.534728 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.534760 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.534767 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.534780 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.534788 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:43Z","lastTransitionTime":"2025-12-04T03:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.637895 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.637953 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.637964 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.637977 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.637985 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:43Z","lastTransitionTime":"2025-12-04T03:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.741144 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.741177 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.741185 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.741198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.741206 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:43Z","lastTransitionTime":"2025-12-04T03:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.849473 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.849519 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.849530 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.849546 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.849561 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:43Z","lastTransitionTime":"2025-12-04T03:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.952509 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.952582 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.952601 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.952620 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:43 crc kubenswrapper[4806]: I1204 03:55:43.952633 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:43Z","lastTransitionTime":"2025-12-04T03:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.055807 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.055873 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.055889 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.055912 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.055958 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:44Z","lastTransitionTime":"2025-12-04T03:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.159498 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.159585 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.159596 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.159629 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.159639 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:44Z","lastTransitionTime":"2025-12-04T03:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.262651 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.262699 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.262708 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.262728 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.262740 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:44Z","lastTransitionTime":"2025-12-04T03:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.365310 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.365359 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.365375 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.365391 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.365401 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:44Z","lastTransitionTime":"2025-12-04T03:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.467838 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.467916 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.467950 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.467985 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.467999 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:44Z","lastTransitionTime":"2025-12-04T03:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.570736 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.570790 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.570801 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.570831 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.570843 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:44Z","lastTransitionTime":"2025-12-04T03:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.673409 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.673468 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.673486 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.673507 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.673522 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:44Z","lastTransitionTime":"2025-12-04T03:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.776106 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.776145 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.776156 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.776171 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.776182 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:44Z","lastTransitionTime":"2025-12-04T03:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.879155 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.879191 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.879199 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.879213 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.879223 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:44Z","lastTransitionTime":"2025-12-04T03:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.982296 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.982367 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.982383 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.982401 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:44 crc kubenswrapper[4806]: I1204 03:55:44.982711 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:44Z","lastTransitionTime":"2025-12-04T03:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.085560 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.085597 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.085609 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.085625 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.085636 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:45Z","lastTransitionTime":"2025-12-04T03:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.188427 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.188473 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.188482 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.188497 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.188510 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:45Z","lastTransitionTime":"2025-12-04T03:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.291411 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.291465 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.291476 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.291495 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.291507 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:45Z","lastTransitionTime":"2025-12-04T03:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.393998 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.394029 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.394037 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.394049 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.394059 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:45Z","lastTransitionTime":"2025-12-04T03:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.423413 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.423467 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.423471 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:45 crc kubenswrapper[4806]: E1204 03:55:45.423561 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.423645 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:45 crc kubenswrapper[4806]: E1204 03:55:45.423729 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:45 crc kubenswrapper[4806]: E1204 03:55:45.423822 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:45 crc kubenswrapper[4806]: E1204 03:55:45.424013 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.443141 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.464631 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:38Z\\\",\\\"message\\\":\\\"1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 03:55:38.189705 6360 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1204 03:55:38.189692 6360 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 03:55:38.189728 6360 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 03:55:38.189313 6360 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1204 03:55:38.189731 6360 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Tim\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.477775 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b9khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41fdb600-3868-48ab-8396-17c2382b6168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b9khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.491826 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.496828 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.496970 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.497045 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.497140 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.497223 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:45Z","lastTransitionTime":"2025-12-04T03:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.505643 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.522402 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"186beef2-bf40-41df-ad1d-870b99e5b6f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59760c3aea1ddf4bdfeda36dd7e6e20bb56985e39f27864475a064d0061325de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd1df0baab4bc2ca707440eaa20c66a5b003a783d48ac8ed0c66d185a6e4d116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4e4be31ada4276e640bcf04ab82968c378cfaa920bc82ce54d925d457c51585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.538661 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.554944 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.570768 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.587219 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.599369 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.599617 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.599704 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.599814 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.599901 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:45Z","lastTransitionTime":"2025-12-04T03:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.603044 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.615188 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.628489 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.641657 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.655488 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.667669 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.683106 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:45Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.704544 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.704582 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.704593 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.704607 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.704618 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:45Z","lastTransitionTime":"2025-12-04T03:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.807355 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.807399 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.807409 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.807425 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.807437 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:45Z","lastTransitionTime":"2025-12-04T03:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.910617 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.910684 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.910702 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.910723 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:45 crc kubenswrapper[4806]: I1204 03:55:45.910741 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:45Z","lastTransitionTime":"2025-12-04T03:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.016824 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.016884 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.016895 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.016918 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.016957 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:46Z","lastTransitionTime":"2025-12-04T03:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.119565 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.119622 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.119637 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.119665 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.119682 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:46Z","lastTransitionTime":"2025-12-04T03:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.224828 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.224880 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.224900 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.224941 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.224956 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:46Z","lastTransitionTime":"2025-12-04T03:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.327833 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.327902 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.327947 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.327974 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.327994 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:46Z","lastTransitionTime":"2025-12-04T03:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.430899 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.430969 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.430981 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.431000 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.431010 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:46Z","lastTransitionTime":"2025-12-04T03:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.534063 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.534141 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.534154 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.534179 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.534193 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:46Z","lastTransitionTime":"2025-12-04T03:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.639469 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.639667 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.639690 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.639716 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.639733 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:46Z","lastTransitionTime":"2025-12-04T03:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.742658 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.743007 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.743031 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.743056 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.743072 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:46Z","lastTransitionTime":"2025-12-04T03:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.845168 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.845227 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.845244 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.845267 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.845282 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:46Z","lastTransitionTime":"2025-12-04T03:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.947894 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.947976 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.947988 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.948007 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:46 crc kubenswrapper[4806]: I1204 03:55:46.948020 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:46Z","lastTransitionTime":"2025-12-04T03:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.050613 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.050664 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.050681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.050701 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.050714 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:47Z","lastTransitionTime":"2025-12-04T03:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.153343 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.153391 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.153406 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.153427 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.153444 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:47Z","lastTransitionTime":"2025-12-04T03:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.256152 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.256191 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.256200 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.256215 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.256226 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:47Z","lastTransitionTime":"2025-12-04T03:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.358817 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.358862 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.358874 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.358891 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.358900 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:47Z","lastTransitionTime":"2025-12-04T03:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.423218 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.423318 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.423269 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:47 crc kubenswrapper[4806]: E1204 03:55:47.423444 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:47 crc kubenswrapper[4806]: E1204 03:55:47.423560 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:47 crc kubenswrapper[4806]: E1204 03:55:47.423635 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.423773 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:47 crc kubenswrapper[4806]: E1204 03:55:47.423829 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.461466 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.461505 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.461519 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.461538 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.461551 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:47Z","lastTransitionTime":"2025-12-04T03:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.564699 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.564751 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.564766 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.564785 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.564800 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:47Z","lastTransitionTime":"2025-12-04T03:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.635176 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.635212 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.635223 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.635238 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.635250 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:47Z","lastTransitionTime":"2025-12-04T03:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:47 crc kubenswrapper[4806]: E1204 03:55:47.651711 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:47Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.655504 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.655531 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.655539 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.655552 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.655562 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:47Z","lastTransitionTime":"2025-12-04T03:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:47 crc kubenswrapper[4806]: E1204 03:55:47.679050 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:47Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.683815 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.683861 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.683958 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.683982 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.683995 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:47Z","lastTransitionTime":"2025-12-04T03:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:47 crc kubenswrapper[4806]: E1204 03:55:47.702621 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:47Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.707622 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.707695 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.707716 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.707744 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.707763 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:47Z","lastTransitionTime":"2025-12-04T03:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:47 crc kubenswrapper[4806]: E1204 03:55:47.724161 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:47Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.729495 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.729548 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.729559 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.729578 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.729619 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:47Z","lastTransitionTime":"2025-12-04T03:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:47 crc kubenswrapper[4806]: E1204 03:55:47.744401 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:47Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:47 crc kubenswrapper[4806]: E1204 03:55:47.744550 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.746737 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.746794 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.746806 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.746829 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.746844 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:47Z","lastTransitionTime":"2025-12-04T03:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.850458 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.850547 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.850559 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.850588 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.850621 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:47Z","lastTransitionTime":"2025-12-04T03:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.954586 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.954667 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.954701 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.954723 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:47 crc kubenswrapper[4806]: I1204 03:55:47.954737 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:47Z","lastTransitionTime":"2025-12-04T03:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.065334 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.065386 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.065403 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.065426 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.065445 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:48Z","lastTransitionTime":"2025-12-04T03:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.168786 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.168851 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.168863 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.168884 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.168899 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:48Z","lastTransitionTime":"2025-12-04T03:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.271470 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.271530 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.271540 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.271559 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.271573 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:48Z","lastTransitionTime":"2025-12-04T03:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.374479 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.374547 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.374561 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.374589 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.374605 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:48Z","lastTransitionTime":"2025-12-04T03:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.477997 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.478068 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.478081 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.478104 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.478119 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:48Z","lastTransitionTime":"2025-12-04T03:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.580708 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.580767 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.580783 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.580802 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.580817 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:48Z","lastTransitionTime":"2025-12-04T03:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.683945 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.683985 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.683997 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.684014 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.684027 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:48Z","lastTransitionTime":"2025-12-04T03:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.786664 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.786713 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.786725 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.786741 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.786751 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:48Z","lastTransitionTime":"2025-12-04T03:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.889164 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.889216 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.889227 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.889248 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.889261 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:48Z","lastTransitionTime":"2025-12-04T03:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.992245 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.992298 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.992307 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.992323 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:48 crc kubenswrapper[4806]: I1204 03:55:48.992334 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:48Z","lastTransitionTime":"2025-12-04T03:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.094979 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.095019 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.095029 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.095044 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.095054 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:49Z","lastTransitionTime":"2025-12-04T03:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.197531 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.197575 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.197586 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.197600 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.197608 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:49Z","lastTransitionTime":"2025-12-04T03:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.300366 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.300411 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.300420 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.300434 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.300444 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:49Z","lastTransitionTime":"2025-12-04T03:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.402993 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.403028 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.403037 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.403051 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.403060 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:49Z","lastTransitionTime":"2025-12-04T03:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.423727 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.423809 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.423833 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:49 crc kubenswrapper[4806]: E1204 03:55:49.423886 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:49 crc kubenswrapper[4806]: E1204 03:55:49.423991 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.423716 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:49 crc kubenswrapper[4806]: E1204 03:55:49.424135 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:49 crc kubenswrapper[4806]: E1204 03:55:49.424237 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.505561 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.505650 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.505664 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.505681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.505692 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:49Z","lastTransitionTime":"2025-12-04T03:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.608725 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.608786 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.608807 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.608830 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.608845 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:49Z","lastTransitionTime":"2025-12-04T03:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.712005 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.712051 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.712061 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.712075 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.712086 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:49Z","lastTransitionTime":"2025-12-04T03:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.814987 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.815053 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.815065 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.815088 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.815102 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:49Z","lastTransitionTime":"2025-12-04T03:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.917863 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.917905 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.917915 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.917945 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:49 crc kubenswrapper[4806]: I1204 03:55:49.917955 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:49Z","lastTransitionTime":"2025-12-04T03:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.020416 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.020482 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.020495 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.020517 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.020530 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:50Z","lastTransitionTime":"2025-12-04T03:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.122815 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.122864 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.122875 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.122891 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.122903 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:50Z","lastTransitionTime":"2025-12-04T03:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.225392 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.225456 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.225467 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.225486 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.225499 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:50Z","lastTransitionTime":"2025-12-04T03:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.328205 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.328262 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.328271 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.328286 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.328302 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:50Z","lastTransitionTime":"2025-12-04T03:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.423535 4806 scope.go:117] "RemoveContainer" containerID="d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4" Dec 04 03:55:50 crc kubenswrapper[4806]: E1204 03:55:50.423687 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.430777 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.430824 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.430834 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.430848 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.430858 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:50Z","lastTransitionTime":"2025-12-04T03:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.532489 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.532526 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.532535 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.532547 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.532558 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:50Z","lastTransitionTime":"2025-12-04T03:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.635194 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.635233 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.635245 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.635261 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.635272 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:50Z","lastTransitionTime":"2025-12-04T03:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.737804 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.737835 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.737843 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.737856 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.737866 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:50Z","lastTransitionTime":"2025-12-04T03:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.841256 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.841298 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.841308 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.841325 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.841335 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:50Z","lastTransitionTime":"2025-12-04T03:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.944137 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.944175 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.944188 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.944204 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:50 crc kubenswrapper[4806]: I1204 03:55:50.944219 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:50Z","lastTransitionTime":"2025-12-04T03:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.047198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.047239 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.047251 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.047269 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.047281 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:51Z","lastTransitionTime":"2025-12-04T03:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.149636 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.149685 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.149696 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.149717 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.149729 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:51Z","lastTransitionTime":"2025-12-04T03:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.253056 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.253109 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.253126 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.253151 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.253168 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:51Z","lastTransitionTime":"2025-12-04T03:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.355984 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.356031 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.356042 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.356062 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.356074 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:51Z","lastTransitionTime":"2025-12-04T03:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.424084 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.424157 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.424099 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:51 crc kubenswrapper[4806]: E1204 03:55:51.424215 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:51 crc kubenswrapper[4806]: E1204 03:55:51.424310 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.424356 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:51 crc kubenswrapper[4806]: E1204 03:55:51.424409 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:51 crc kubenswrapper[4806]: E1204 03:55:51.424469 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.459578 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.459623 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.459637 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.459655 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.459668 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:51Z","lastTransitionTime":"2025-12-04T03:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.561903 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.561953 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.561966 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.561983 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.561995 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:51Z","lastTransitionTime":"2025-12-04T03:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.661660 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs\") pod \"network-metrics-daemon-b9khd\" (UID: \"41fdb600-3868-48ab-8396-17c2382b6168\") " pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:51 crc kubenswrapper[4806]: E1204 03:55:51.661868 4806 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 03:55:51 crc kubenswrapper[4806]: E1204 03:55:51.661986 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs podName:41fdb600-3868-48ab-8396-17c2382b6168 nodeName:}" failed. No retries permitted until 2025-12-04 03:56:23.661961537 +0000 UTC m=+98.520474545 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs") pod "network-metrics-daemon-b9khd" (UID: "41fdb600-3868-48ab-8396-17c2382b6168") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.663761 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.663797 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.663808 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.663823 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.663834 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:51Z","lastTransitionTime":"2025-12-04T03:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.765745 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.765820 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.765830 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.765844 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.765854 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:51Z","lastTransitionTime":"2025-12-04T03:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.868848 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.868889 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.868899 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.868914 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.868940 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:51Z","lastTransitionTime":"2025-12-04T03:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.970793 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.970828 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.970838 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.970855 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:51 crc kubenswrapper[4806]: I1204 03:55:51.970867 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:51Z","lastTransitionTime":"2025-12-04T03:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.073038 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.073085 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.073095 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.073111 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.073122 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:52Z","lastTransitionTime":"2025-12-04T03:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.176058 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.176102 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.176112 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.176128 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.176141 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:52Z","lastTransitionTime":"2025-12-04T03:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.279344 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.279698 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.279777 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.279862 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.279950 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:52Z","lastTransitionTime":"2025-12-04T03:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.382651 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.382696 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.382706 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.382719 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.382728 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:52Z","lastTransitionTime":"2025-12-04T03:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.436313 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.485167 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.485233 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.485245 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.485264 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.485275 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:52Z","lastTransitionTime":"2025-12-04T03:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.587616 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.587665 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.587673 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.587687 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.587696 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:52Z","lastTransitionTime":"2025-12-04T03:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.690477 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.690534 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.690553 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.690577 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.690593 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:52Z","lastTransitionTime":"2025-12-04T03:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.770546 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dsclk_4f13bcfc-c287-40fd-b8f6-e4bbea8c6577/kube-multus/0.log" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.770602 4806 generic.go:334] "Generic (PLEG): container finished" podID="4f13bcfc-c287-40fd-b8f6-e4bbea8c6577" containerID="49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b" exitCode=1 Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.770699 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dsclk" event={"ID":"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577","Type":"ContainerDied","Data":"49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b"} Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.771130 4806 scope.go:117] "RemoveContainer" containerID="49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.784475 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:52Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.797345 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.797405 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.797417 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.797433 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.797446 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:52Z","lastTransitionTime":"2025-12-04T03:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.805739 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:52Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.818340 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:52Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.832495 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:52Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.850689 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:52Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.862584 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6649e412-5f28-4771-9d38-db928b07e8a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d42708d1f2974b0a6eb579a33c340969fa6df1100355754cb12ceb117f305f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:52Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.876198 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"186beef2-bf40-41df-ad1d-870b99e5b6f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59760c3aea1ddf4bdfeda36dd7e6e20bb56985e39f27864475a064d0061325de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd1df0baab4bc2ca707440eaa20c66a5b003a783d48ac8ed0c66d185a6e4d116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4e4be31ada4276e640bcf04ab82968c378cfaa920bc82ce54d925d457c51585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:52Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.888776 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:52Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.901015 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.901063 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.901075 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.901093 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.901106 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:52Z","lastTransitionTime":"2025-12-04T03:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.906001 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:52Z\\\",\\\"message\\\":\\\"2025-12-04T03:55:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00ac28ad-f9e0-4f82-9235-011a62af4d92\\\\n2025-12-04T03:55:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00ac28ad-f9e0-4f82-9235-011a62af4d92 to /host/opt/cni/bin/\\\\n2025-12-04T03:55:06Z [verbose] multus-daemon started\\\\n2025-12-04T03:55:06Z [verbose] Readiness Indicator file check\\\\n2025-12-04T03:55:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:52Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.917106 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:52Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.928392 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:52Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.941044 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:52Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.952556 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:52Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.970176 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:38Z\\\",\\\"message\\\":\\\"1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 03:55:38.189705 6360 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1204 03:55:38.189692 6360 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 03:55:38.189728 6360 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 03:55:38.189313 6360 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1204 03:55:38.189731 6360 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Tim\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:52Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.980103 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b9khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41fdb600-3868-48ab-8396-17c2382b6168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b9khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:52Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:52 crc kubenswrapper[4806]: I1204 03:55:52.991668 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:52Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.003015 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.003076 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.003086 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.003100 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.003111 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:53Z","lastTransitionTime":"2025-12-04T03:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.004177 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.014419 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.105585 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.105637 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.105650 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.105668 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.105679 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:53Z","lastTransitionTime":"2025-12-04T03:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.208413 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.208905 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.208941 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.208962 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.208974 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:53Z","lastTransitionTime":"2025-12-04T03:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.311642 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.311678 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.311687 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.311701 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.311712 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:53Z","lastTransitionTime":"2025-12-04T03:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.413566 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.413603 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.413612 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.413625 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.413636 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:53Z","lastTransitionTime":"2025-12-04T03:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.423264 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.423291 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:53 crc kubenswrapper[4806]: E1204 03:55:53.423352 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.423511 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.423516 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:53 crc kubenswrapper[4806]: E1204 03:55:53.423507 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:53 crc kubenswrapper[4806]: E1204 03:55:53.423558 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:53 crc kubenswrapper[4806]: E1204 03:55:53.423603 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.515877 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.515952 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.515962 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.515978 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.515994 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:53Z","lastTransitionTime":"2025-12-04T03:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.618602 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.618635 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.618652 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.618670 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.618682 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:53Z","lastTransitionTime":"2025-12-04T03:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.722053 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.722109 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.722120 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.722142 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.722155 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:53Z","lastTransitionTime":"2025-12-04T03:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.777703 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dsclk_4f13bcfc-c287-40fd-b8f6-e4bbea8c6577/kube-multus/0.log" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.777799 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dsclk" event={"ID":"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577","Type":"ContainerStarted","Data":"a6e29eb2e4c712cf0fcf618844fe548f0c38982e08169a900c816f7a9a6b71ba"} Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.794286 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.808120 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e29eb2e4c712cf0fcf618844fe548f0c38982e08169a900c816f7a9a6b71ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:52Z\\\",\\\"message\\\":\\\"2025-12-04T03:55:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00ac28ad-f9e0-4f82-9235-011a62af4d92\\\\n2025-12-04T03:55:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00ac28ad-f9e0-4f82-9235-011a62af4d92 to /host/opt/cni/bin/\\\\n2025-12-04T03:55:06Z [verbose] multus-daemon started\\\\n2025-12-04T03:55:06Z [verbose] Readiness Indicator file check\\\\n2025-12-04T03:55:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.820426 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.824141 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.824183 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.824198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.824218 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.824231 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:53Z","lastTransitionTime":"2025-12-04T03:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.831597 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.844855 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.857442 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.877912 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:38Z\\\",\\\"message\\\":\\\"1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 03:55:38.189705 6360 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1204 03:55:38.189692 6360 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 03:55:38.189728 6360 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 03:55:38.189313 6360 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1204 03:55:38.189731 6360 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Tim\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.888693 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b9khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41fdb600-3868-48ab-8396-17c2382b6168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b9khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.900256 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.918798 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.926520 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.926558 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.926568 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.926584 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.926596 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:53Z","lastTransitionTime":"2025-12-04T03:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.930357 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.941015 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.954418 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.966663 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.981862 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:53 crc kubenswrapper[4806]: I1204 03:55:53.993692 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:53Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.003989 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6649e412-5f28-4771-9d38-db928b07e8a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d42708d1f2974b0a6eb579a33c340969fa6df1100355754cb12ceb117f305f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:54Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.016249 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"186beef2-bf40-41df-ad1d-870b99e5b6f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59760c3aea1ddf4bdfeda36dd7e6e20bb56985e39f27864475a064d0061325de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd1df0baab4bc2ca707440eaa20c66a5b003a783d48ac8ed0c66d185a6e4d116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4e4be31ada4276e640bcf04ab82968c378cfaa920bc82ce54d925d457c51585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:54Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.029592 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.029633 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.029643 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.029659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.029669 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:54Z","lastTransitionTime":"2025-12-04T03:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.131974 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.132040 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.132054 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.132073 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.132084 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:54Z","lastTransitionTime":"2025-12-04T03:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.234360 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.234403 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.234412 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.234427 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.234435 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:54Z","lastTransitionTime":"2025-12-04T03:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.337499 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.337540 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.337550 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.337564 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.337574 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:54Z","lastTransitionTime":"2025-12-04T03:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.440148 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.440197 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.440208 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.440224 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.440236 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:54Z","lastTransitionTime":"2025-12-04T03:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.542299 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.542346 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.542354 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.542371 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.542565 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:54Z","lastTransitionTime":"2025-12-04T03:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.644566 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.644608 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.644620 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.644637 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.644647 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:54Z","lastTransitionTime":"2025-12-04T03:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.746567 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.746614 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.746624 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.746640 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.746652 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:54Z","lastTransitionTime":"2025-12-04T03:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.849203 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.849238 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.849247 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.849259 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.849270 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:54Z","lastTransitionTime":"2025-12-04T03:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.951591 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.951643 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.951653 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.951667 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:54 crc kubenswrapper[4806]: I1204 03:55:54.951675 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:54Z","lastTransitionTime":"2025-12-04T03:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.053815 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.053863 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.053877 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.053893 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.053904 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:55Z","lastTransitionTime":"2025-12-04T03:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.156662 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.156697 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.156704 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.156717 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.156731 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:55Z","lastTransitionTime":"2025-12-04T03:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.259161 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.259212 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.259229 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.259247 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.259259 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:55Z","lastTransitionTime":"2025-12-04T03:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.361595 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.361636 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.361645 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.361676 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.361686 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:55Z","lastTransitionTime":"2025-12-04T03:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.422691 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.422691 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:55 crc kubenswrapper[4806]: E1204 03:55:55.422817 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.422715 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:55 crc kubenswrapper[4806]: E1204 03:55:55.422895 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.422691 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:55 crc kubenswrapper[4806]: E1204 03:55:55.422979 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:55 crc kubenswrapper[4806]: E1204 03:55:55.423033 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.434488 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.444004 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.456012 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.463888 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.463916 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.463943 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.463958 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.463968 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:55Z","lastTransitionTime":"2025-12-04T03:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.466245 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.482842 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:38Z\\\",\\\"message\\\":\\\"1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 03:55:38.189705 6360 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1204 03:55:38.189692 6360 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 03:55:38.189728 6360 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 03:55:38.189313 6360 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1204 03:55:38.189731 6360 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Tim\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.493052 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b9khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41fdb600-3868-48ab-8396-17c2382b6168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b9khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.503797 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.514668 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.525025 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"186beef2-bf40-41df-ad1d-870b99e5b6f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59760c3aea1ddf4bdfeda36dd7e6e20bb56985e39f27864475a064d0061325de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd1df0baab4bc2ca707440eaa20c66a5b003a783d48ac8ed0c66d185a6e4d116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4e4be31ada4276e640bcf04ab82968c378cfaa920bc82ce54d925d457c51585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.539154 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.565902 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.565958 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.565968 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.565982 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.565993 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:55Z","lastTransitionTime":"2025-12-04T03:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.566295 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.577115 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.592174 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.605913 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.616615 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6649e412-5f28-4771-9d38-db928b07e8a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d42708d1f2974b0a6eb579a33c340969fa6df1100355754cb12ceb117f305f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.626576 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.639091 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.656152 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e29eb2e4c712cf0fcf618844fe548f0c38982e08169a900c816f7a9a6b71ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:52Z\\\",\\\"message\\\":\\\"2025-12-04T03:55:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00ac28ad-f9e0-4f82-9235-011a62af4d92\\\\n2025-12-04T03:55:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00ac28ad-f9e0-4f82-9235-011a62af4d92 to /host/opt/cni/bin/\\\\n2025-12-04T03:55:06Z [verbose] multus-daemon started\\\\n2025-12-04T03:55:06Z [verbose] Readiness Indicator file check\\\\n2025-12-04T03:55:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:55Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.667372 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.667566 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.667632 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.667693 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.667752 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:55Z","lastTransitionTime":"2025-12-04T03:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.770200 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.770269 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.770294 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.770318 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.770336 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:55Z","lastTransitionTime":"2025-12-04T03:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.873029 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.873064 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.873073 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.873086 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.873096 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:55Z","lastTransitionTime":"2025-12-04T03:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.975317 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.975350 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.975358 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.975371 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:55 crc kubenswrapper[4806]: I1204 03:55:55.975379 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:55Z","lastTransitionTime":"2025-12-04T03:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.077838 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.077875 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.077884 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.077897 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.077906 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:56Z","lastTransitionTime":"2025-12-04T03:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.180002 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.180319 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.180411 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.180496 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.180578 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:56Z","lastTransitionTime":"2025-12-04T03:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.283044 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.283074 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.283083 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.283094 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.283103 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:56Z","lastTransitionTime":"2025-12-04T03:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.384845 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.385130 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.385248 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.385317 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.385382 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:56Z","lastTransitionTime":"2025-12-04T03:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.488162 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.488414 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.488511 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.488773 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.488830 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:56Z","lastTransitionTime":"2025-12-04T03:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.590826 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.590869 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.590881 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.590899 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.590916 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:56Z","lastTransitionTime":"2025-12-04T03:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.693088 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.693117 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.693126 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.693140 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.693150 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:56Z","lastTransitionTime":"2025-12-04T03:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.795412 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.795470 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.795482 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.795516 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.795527 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:56Z","lastTransitionTime":"2025-12-04T03:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.897592 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.897623 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.897633 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.897650 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.897660 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:56Z","lastTransitionTime":"2025-12-04T03:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.999769 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.999802 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.999810 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.999827 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:56 crc kubenswrapper[4806]: I1204 03:55:56.999837 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:56Z","lastTransitionTime":"2025-12-04T03:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.102094 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.102140 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.102151 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.102167 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.102177 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:57Z","lastTransitionTime":"2025-12-04T03:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.204466 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.204497 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.204507 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.204521 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.204531 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:57Z","lastTransitionTime":"2025-12-04T03:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.307264 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.307309 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.307322 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.307339 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.307351 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:57Z","lastTransitionTime":"2025-12-04T03:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.416774 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.416842 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.416863 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.416887 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.416898 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:57Z","lastTransitionTime":"2025-12-04T03:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.424738 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:57 crc kubenswrapper[4806]: E1204 03:55:57.424869 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.425089 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:57 crc kubenswrapper[4806]: E1204 03:55:57.425149 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.425268 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:57 crc kubenswrapper[4806]: E1204 03:55:57.425332 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.425453 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:57 crc kubenswrapper[4806]: E1204 03:55:57.425819 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.520120 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.520162 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.520173 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.520191 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.520202 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:57Z","lastTransitionTime":"2025-12-04T03:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.623229 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.623290 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.623304 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.623327 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.623341 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:57Z","lastTransitionTime":"2025-12-04T03:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.726917 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.727003 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.727016 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.727040 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.727055 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:57Z","lastTransitionTime":"2025-12-04T03:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.829545 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.829589 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.829601 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.829616 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.829628 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:57Z","lastTransitionTime":"2025-12-04T03:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.847284 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.847344 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.847378 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.847400 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.847412 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:57Z","lastTransitionTime":"2025-12-04T03:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:57 crc kubenswrapper[4806]: E1204 03:55:57.861345 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:57Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.865207 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.865238 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.865248 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.865263 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.865273 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:57Z","lastTransitionTime":"2025-12-04T03:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:57 crc kubenswrapper[4806]: E1204 03:55:57.877840 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:57Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.882057 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.882125 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.882140 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.882182 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.882196 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:57Z","lastTransitionTime":"2025-12-04T03:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:57 crc kubenswrapper[4806]: E1204 03:55:57.895442 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:57Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.899204 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.899244 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.899255 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.899272 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.899286 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:57Z","lastTransitionTime":"2025-12-04T03:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:57 crc kubenswrapper[4806]: E1204 03:55:57.913906 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:57Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.920784 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.920823 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.920835 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.920850 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.920862 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:57Z","lastTransitionTime":"2025-12-04T03:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:57 crc kubenswrapper[4806]: E1204 03:55:57.933593 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:55:57Z is after 2025-08-24T17:21:41Z" Dec 04 03:55:57 crc kubenswrapper[4806]: E1204 03:55:57.933765 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.935515 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.935579 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.935593 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.935612 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:57 crc kubenswrapper[4806]: I1204 03:55:57.935627 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:57Z","lastTransitionTime":"2025-12-04T03:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.037796 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.037838 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.037851 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.037866 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.037879 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:58Z","lastTransitionTime":"2025-12-04T03:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.140277 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.140313 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.140321 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.140335 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.140344 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:58Z","lastTransitionTime":"2025-12-04T03:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.243349 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.243428 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.243442 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.243464 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.243481 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:58Z","lastTransitionTime":"2025-12-04T03:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.346377 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.346426 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.346439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.346454 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.346463 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:58Z","lastTransitionTime":"2025-12-04T03:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.448838 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.448916 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.448956 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.448972 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.448981 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:58Z","lastTransitionTime":"2025-12-04T03:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.551156 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.551223 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.551236 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.551254 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.551265 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:58Z","lastTransitionTime":"2025-12-04T03:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.653866 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.653916 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.653956 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.653970 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.653977 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:58Z","lastTransitionTime":"2025-12-04T03:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.756758 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.756812 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.756825 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.756840 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.756850 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:58Z","lastTransitionTime":"2025-12-04T03:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.860314 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.860353 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.860361 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.860379 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.860388 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:58Z","lastTransitionTime":"2025-12-04T03:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.963154 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.963202 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.963214 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.963228 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:58 crc kubenswrapper[4806]: I1204 03:55:58.963238 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:58Z","lastTransitionTime":"2025-12-04T03:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.070498 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.070560 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.070575 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.070596 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.070610 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:59Z","lastTransitionTime":"2025-12-04T03:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.172519 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.172554 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.172561 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.172573 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.172582 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:59Z","lastTransitionTime":"2025-12-04T03:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.274521 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.274560 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.274572 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.274587 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.274630 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:59Z","lastTransitionTime":"2025-12-04T03:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.377016 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.377060 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.377068 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.377082 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.377093 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:59Z","lastTransitionTime":"2025-12-04T03:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.424977 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.425146 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:55:59 crc kubenswrapper[4806]: E1204 03:55:59.425154 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:55:59 crc kubenswrapper[4806]: E1204 03:55:59.425323 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.423521 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:55:59 crc kubenswrapper[4806]: E1204 03:55:59.425533 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.425640 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:55:59 crc kubenswrapper[4806]: E1204 03:55:59.425749 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.479063 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.479186 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.479197 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.479210 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.479224 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:59Z","lastTransitionTime":"2025-12-04T03:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.581905 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.581997 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.582007 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.582021 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.582034 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:59Z","lastTransitionTime":"2025-12-04T03:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.684679 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.684723 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.684735 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.684751 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.684762 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:59Z","lastTransitionTime":"2025-12-04T03:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.787998 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.788042 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.788051 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.788065 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.788075 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:59Z","lastTransitionTime":"2025-12-04T03:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.890072 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.890104 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.890112 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.890125 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.890134 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:59Z","lastTransitionTime":"2025-12-04T03:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.992047 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.992079 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.992089 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.992103 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:55:59 crc kubenswrapper[4806]: I1204 03:55:59.992115 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:55:59Z","lastTransitionTime":"2025-12-04T03:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.094672 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.094786 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.094799 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.094816 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.094825 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:00Z","lastTransitionTime":"2025-12-04T03:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.196638 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.196672 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.196681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.196695 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.196704 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:00Z","lastTransitionTime":"2025-12-04T03:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.299190 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.299302 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.299316 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.299330 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.299359 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:00Z","lastTransitionTime":"2025-12-04T03:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.401389 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.401431 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.401446 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.401465 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.401477 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:00Z","lastTransitionTime":"2025-12-04T03:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.504431 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.504476 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.504486 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.504501 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.504511 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:00Z","lastTransitionTime":"2025-12-04T03:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.607048 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.607118 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.607140 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.607159 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.607177 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:00Z","lastTransitionTime":"2025-12-04T03:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.708854 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.708898 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.708911 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.708959 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.708972 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:00Z","lastTransitionTime":"2025-12-04T03:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.811512 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.812200 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.812333 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.812437 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.812503 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:00Z","lastTransitionTime":"2025-12-04T03:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.914768 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.914825 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.914840 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.914860 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:00 crc kubenswrapper[4806]: I1204 03:56:00.914872 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:00Z","lastTransitionTime":"2025-12-04T03:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.016705 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.017077 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.017224 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.017324 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.017402 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:01Z","lastTransitionTime":"2025-12-04T03:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.119549 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.119595 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.119604 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.119619 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.119632 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:01Z","lastTransitionTime":"2025-12-04T03:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.222100 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.222167 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.222184 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.222202 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.222215 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:01Z","lastTransitionTime":"2025-12-04T03:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.325559 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.326042 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.326238 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.326483 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.326697 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:01Z","lastTransitionTime":"2025-12-04T03:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.423481 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.423577 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:01 crc kubenswrapper[4806]: E1204 03:56:01.423611 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.423640 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.423640 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:01 crc kubenswrapper[4806]: E1204 03:56:01.423728 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:01 crc kubenswrapper[4806]: E1204 03:56:01.423747 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:01 crc kubenswrapper[4806]: E1204 03:56:01.424125 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.424371 4806 scope.go:117] "RemoveContainer" containerID="d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.428915 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.428963 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.428972 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.428985 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.428995 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:01Z","lastTransitionTime":"2025-12-04T03:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.531526 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.531570 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.531582 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.531599 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.531612 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:01Z","lastTransitionTime":"2025-12-04T03:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.634077 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.634111 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.634120 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.634134 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.634142 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:01Z","lastTransitionTime":"2025-12-04T03:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.736737 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.736802 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.736816 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.736855 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.736868 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:01Z","lastTransitionTime":"2025-12-04T03:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.802093 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/2.log" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.806831 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerStarted","Data":"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d"} Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.807835 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.819272 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:01Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.831573 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:01Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.839321 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.839350 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.839360 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.839376 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.839386 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:01Z","lastTransitionTime":"2025-12-04T03:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.845213 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e29eb2e4c712cf0fcf618844fe548f0c38982e08169a900c816f7a9a6b71ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:52Z\\\",\\\"message\\\":\\\"2025-12-04T03:55:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00ac28ad-f9e0-4f82-9235-011a62af4d92\\\\n2025-12-04T03:55:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00ac28ad-f9e0-4f82-9235-011a62af4d92 to /host/opt/cni/bin/\\\\n2025-12-04T03:55:06Z [verbose] multus-daemon started\\\\n2025-12-04T03:55:06Z [verbose] Readiness Indicator file check\\\\n2025-12-04T03:55:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:01Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.857771 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:01Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.866697 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:01Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.879263 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:01Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.892788 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:01Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.911642 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:38Z\\\",\\\"message\\\":\\\"1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 03:55:38.189705 6360 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1204 03:55:38.189692 6360 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 03:55:38.189728 6360 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 03:55:38.189313 6360 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1204 03:55:38.189731 6360 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Tim\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:01Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.926972 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b9khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41fdb600-3868-48ab-8396-17c2382b6168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b9khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:01Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.942131 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.942187 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.942198 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.942219 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.942232 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:01Z","lastTransitionTime":"2025-12-04T03:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.943580 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:01Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.959525 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:01Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.977683 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"186beef2-bf40-41df-ad1d-870b99e5b6f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59760c3aea1ddf4bdfeda36dd7e6e20bb56985e39f27864475a064d0061325de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd1df0baab4bc2ca707440eaa20c66a5b003a783d48ac8ed0c66d185a6e4d116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4e4be31ada4276e640bcf04ab82968c378cfaa920bc82ce54d925d457c51585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:01Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:01 crc kubenswrapper[4806]: I1204 03:56:01.995410 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:01Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.013152 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:02Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.048472 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.048514 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.048545 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.048561 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.048574 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:02Z","lastTransitionTime":"2025-12-04T03:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.051461 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:02Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.068946 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:02Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.084807 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:02Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.096527 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6649e412-5f28-4771-9d38-db928b07e8a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d42708d1f2974b0a6eb579a33c340969fa6df1100355754cb12ceb117f305f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:02Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.152049 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.152109 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.152120 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.152137 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.152147 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:02Z","lastTransitionTime":"2025-12-04T03:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.255199 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.255239 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.255252 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.255269 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.255281 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:02Z","lastTransitionTime":"2025-12-04T03:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.358145 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.358179 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.358188 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.358204 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.358213 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:02Z","lastTransitionTime":"2025-12-04T03:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.461324 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.461400 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.461415 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.461431 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.461444 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:02Z","lastTransitionTime":"2025-12-04T03:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.563775 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.563814 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.563824 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.563838 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.563848 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:02Z","lastTransitionTime":"2025-12-04T03:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.666277 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.666353 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.666374 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.666399 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.666422 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:02Z","lastTransitionTime":"2025-12-04T03:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.769604 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.769687 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.769709 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.769736 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.769759 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:02Z","lastTransitionTime":"2025-12-04T03:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.812757 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/3.log" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.813499 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/2.log" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.818402 4806 generic.go:334] "Generic (PLEG): container finished" podID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerID="7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d" exitCode=1 Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.818464 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerDied","Data":"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d"} Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.818521 4806 scope.go:117] "RemoveContainer" containerID="d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.819760 4806 scope.go:117] "RemoveContainer" containerID="7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d" Dec 04 03:56:02 crc kubenswrapper[4806]: E1204 03:56:02.820182 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.836202 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:02Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.856894 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e29eb2e4c712cf0fcf618844fe548f0c38982e08169a900c816f7a9a6b71ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:52Z\\\",\\\"message\\\":\\\"2025-12-04T03:55:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00ac28ad-f9e0-4f82-9235-011a62af4d92\\\\n2025-12-04T03:55:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00ac28ad-f9e0-4f82-9235-011a62af4d92 to /host/opt/cni/bin/\\\\n2025-12-04T03:55:06Z [verbose] multus-daemon started\\\\n2025-12-04T03:55:06Z [verbose] Readiness Indicator file check\\\\n2025-12-04T03:55:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:02Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.873458 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:02Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.874186 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.874218 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.874229 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.874245 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.874255 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:02Z","lastTransitionTime":"2025-12-04T03:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.898893 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:02Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.913575 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:02Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.925175 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:02Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.938703 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:02Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.951174 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:02Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.961801 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:02Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.976641 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.976691 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.976706 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.976727 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:02 crc kubenswrapper[4806]: I1204 03:56:02.976741 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:02Z","lastTransitionTime":"2025-12-04T03:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.014934 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0ffbe7be541a12dac8d7e29250d3969f960f01391ef741293293a0d3046f3c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:38Z\\\",\\\"message\\\":\\\"1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 03:55:38.189705 6360 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1204 03:55:38.189692 6360 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1204 03:55:38.189728 6360 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1204 03:55:38.189313 6360 ovn.go:134] Ensuring zone local for Pod openshift-kube-apiserver/kube-apiserver-crc in node crc\\\\nI1204 03:55:38.189731 6360 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Tim\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:56:02Z\\\",\\\"message\\\":\\\"leStack,ClusterIPs:[10.217.4.213],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1204 03:56:02.434417 6668 lb_config.go:1031] Cluster endpoints for openshift-console/downloads for network=default are: map[]\\\\nI1204 03:56:02.434426 6668 services_controller.go:443] Built service openshift-console/downloads LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.213\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:80, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1204 03:56:02.434437 6668 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:56:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.028055 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b9khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41fdb600-3868-48ab-8396-17c2382b6168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b9khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.042062 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.054073 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6649e412-5f28-4771-9d38-db928b07e8a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d42708d1f2974b0a6eb579a33c340969fa6df1100355754cb12ceb117f305f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.066913 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"186beef2-bf40-41df-ad1d-870b99e5b6f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59760c3aea1ddf4bdfeda36dd7e6e20bb56985e39f27864475a064d0061325de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd1df0baab4bc2ca707440eaa20c66a5b003a783d48ac8ed0c66d185a6e4d116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4e4be31ada4276e640bcf04ab82968c378cfaa920bc82ce54d925d457c51585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.076937 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.078425 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.078473 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.078485 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.078500 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.078528 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:03Z","lastTransitionTime":"2025-12-04T03:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.088765 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.098157 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.111161 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.180513 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.180561 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.180571 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.180587 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.180599 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:03Z","lastTransitionTime":"2025-12-04T03:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.283315 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.283642 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.283751 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.283845 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.283953 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:03Z","lastTransitionTime":"2025-12-04T03:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.385735 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.385779 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.385794 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.385809 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.385819 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:03Z","lastTransitionTime":"2025-12-04T03:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.423633 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.423708 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.423758 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:03 crc kubenswrapper[4806]: E1204 03:56:03.423803 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.423708 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:03 crc kubenswrapper[4806]: E1204 03:56:03.424008 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:03 crc kubenswrapper[4806]: E1204 03:56:03.424074 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:03 crc kubenswrapper[4806]: E1204 03:56:03.424155 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.436162 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.487981 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.488251 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.488321 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.488415 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.488494 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:03Z","lastTransitionTime":"2025-12-04T03:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.591942 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.591992 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.592004 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.592024 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.592037 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:03Z","lastTransitionTime":"2025-12-04T03:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.694401 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.694447 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.694459 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.694474 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.694487 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:03Z","lastTransitionTime":"2025-12-04T03:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.797716 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.797774 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.797784 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.797802 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.797835 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:03Z","lastTransitionTime":"2025-12-04T03:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.823890 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/3.log" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.830286 4806 scope.go:117] "RemoveContainer" containerID="7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d" Dec 04 03:56:03 crc kubenswrapper[4806]: E1204 03:56:03.830469 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.847891 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.862022 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.874958 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.895780 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:56:02Z\\\",\\\"message\\\":\\\"leStack,ClusterIPs:[10.217.4.213],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1204 03:56:02.434417 6668 lb_config.go:1031] Cluster endpoints for openshift-console/downloads for network=default are: map[]\\\\nI1204 03:56:02.434426 6668 services_controller.go:443] Built service openshift-console/downloads LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.213\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:80, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1204 03:56:02.434437 6668 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:56:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.899575 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.899638 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.899654 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.899673 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.899704 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:03Z","lastTransitionTime":"2025-12-04T03:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.906769 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b9khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41fdb600-3868-48ab-8396-17c2382b6168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b9khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.918321 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.928746 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6649e412-5f28-4771-9d38-db928b07e8a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d42708d1f2974b0a6eb579a33c340969fa6df1100355754cb12ceb117f305f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.943296 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"186beef2-bf40-41df-ad1d-870b99e5b6f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59760c3aea1ddf4bdfeda36dd7e6e20bb56985e39f27864475a064d0061325de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd1df0baab4bc2ca707440eaa20c66a5b003a783d48ac8ed0c66d185a6e4d116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4e4be31ada4276e640bcf04ab82968c378cfaa920bc82ce54d925d457c51585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.954410 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.970664 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.980477 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:03 crc kubenswrapper[4806]: I1204 03:56:03.995310 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:03Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.001912 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.001974 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.001985 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.002000 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.002012 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:04Z","lastTransitionTime":"2025-12-04T03:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.012538 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.027789 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e29eb2e4c712cf0fcf618844fe548f0c38982e08169a900c816f7a9a6b71ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:52Z\\\",\\\"message\\\":\\\"2025-12-04T03:55:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00ac28ad-f9e0-4f82-9235-011a62af4d92\\\\n2025-12-04T03:55:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00ac28ad-f9e0-4f82-9235-011a62af4d92 to /host/opt/cni/bin/\\\\n2025-12-04T03:55:06Z [verbose] multus-daemon started\\\\n2025-12-04T03:55:06Z [verbose] Readiness Indicator file check\\\\n2025-12-04T03:55:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.043917 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.066485 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db305f2-dcdf-4e2c-94ae-0469ce8e27a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cc26e5377fcd914affc21e46397fd6ebe234488cc0eadac1c4e9b543301fca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87e159737687699b84f24ff0e37415ca1728f856572a87b95e5588537606f0ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f2f51aea4a8f01c7d96bc057fef542163504caf5a97320bf382e60eabaec8cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec6de5f42f8d12b7fb0770059775b12e82faad5a81e2c0e9ef85f6212cffae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://25f41c8816799ba262455a943cacc42000e10bcea3a3f90ebca0036e42d5927b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://966a63a12a48455423a3f77e6055886b344e0215cc16c2a2d89bdf1ed967f06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://966a63a12a48455423a3f77e6055886b344e0215cc16c2a2d89bdf1ed967f06d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e851dfa78c781cb95e3e5a3ea73c38d0c4f1059eebace9a6e17b78a2fc879b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e851dfa78c781cb95e3e5a3ea73c38d0c4f1059eebace9a6e17b78a2fc879b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://90ad1f66cf96de520eed60dd21a9817ee3525980fcb43c1f7fd942085c3b49aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ad1f66cf96de520eed60dd21a9817ee3525980fcb43c1f7fd942085c3b49aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.081210 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.097109 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.104443 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.104488 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.104498 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.104514 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.104526 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:04Z","lastTransitionTime":"2025-12-04T03:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.111156 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:04Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.206817 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.206847 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.206856 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.206869 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.206877 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:04Z","lastTransitionTime":"2025-12-04T03:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.309252 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.309297 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.309309 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.309323 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.309332 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:04Z","lastTransitionTime":"2025-12-04T03:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.411087 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.411392 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.411402 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.411417 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.411426 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:04Z","lastTransitionTime":"2025-12-04T03:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.514829 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.514886 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.514900 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.514918 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.514944 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:04Z","lastTransitionTime":"2025-12-04T03:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.617677 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.617757 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.617780 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.617815 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.617837 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:04Z","lastTransitionTime":"2025-12-04T03:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.720912 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.720997 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.721012 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.721029 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.721040 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:04Z","lastTransitionTime":"2025-12-04T03:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.823545 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.823575 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.823582 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.823595 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.823603 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:04Z","lastTransitionTime":"2025-12-04T03:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.925591 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.925634 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.925646 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.925660 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:04 crc kubenswrapper[4806]: I1204 03:56:04.925669 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:04Z","lastTransitionTime":"2025-12-04T03:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.028642 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.028681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.028689 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.028702 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.028710 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:05Z","lastTransitionTime":"2025-12-04T03:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.130975 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.131036 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.131048 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.131063 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.131074 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:05Z","lastTransitionTime":"2025-12-04T03:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.234527 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.234598 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.234619 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.234647 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.234668 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:05Z","lastTransitionTime":"2025-12-04T03:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.337541 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.337581 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.337592 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.337607 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.337618 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:05Z","lastTransitionTime":"2025-12-04T03:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.423626 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:05 crc kubenswrapper[4806]: E1204 03:56:05.423862 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.423877 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.423914 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.424029 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:05 crc kubenswrapper[4806]: E1204 03:56:05.424157 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:05 crc kubenswrapper[4806]: E1204 03:56:05.424311 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:05 crc kubenswrapper[4806]: E1204 03:56:05.424456 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.440199 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.440244 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.440259 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.440280 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.440295 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:05Z","lastTransitionTime":"2025-12-04T03:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.443375 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.458578 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e29eb2e4c712cf0fcf618844fe548f0c38982e08169a900c816f7a9a6b71ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:52Z\\\",\\\"message\\\":\\\"2025-12-04T03:55:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00ac28ad-f9e0-4f82-9235-011a62af4d92\\\\n2025-12-04T03:55:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00ac28ad-f9e0-4f82-9235-011a62af4d92 to /host/opt/cni/bin/\\\\n2025-12-04T03:55:06Z [verbose] multus-daemon started\\\\n2025-12-04T03:55:06Z [verbose] Readiness Indicator file check\\\\n2025-12-04T03:55:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.469054 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.489480 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db305f2-dcdf-4e2c-94ae-0469ce8e27a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cc26e5377fcd914affc21e46397fd6ebe234488cc0eadac1c4e9b543301fca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87e159737687699b84f24ff0e37415ca1728f856572a87b95e5588537606f0ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f2f51aea4a8f01c7d96bc057fef542163504caf5a97320bf382e60eabaec8cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec6de5f42f8d12b7fb0770059775b12e82faad5a81e2c0e9ef85f6212cffae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://25f41c8816799ba262455a943cacc42000e10bcea3a3f90ebca0036e42d5927b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://966a63a12a48455423a3f77e6055886b344e0215cc16c2a2d89bdf1ed967f06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://966a63a12a48455423a3f77e6055886b344e0215cc16c2a2d89bdf1ed967f06d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e851dfa78c781cb95e3e5a3ea73c38d0c4f1059eebace9a6e17b78a2fc879b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e851dfa78c781cb95e3e5a3ea73c38d0c4f1059eebace9a6e17b78a2fc879b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://90ad1f66cf96de520eed60dd21a9817ee3525980fcb43c1f7fd942085c3b49aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ad1f66cf96de520eed60dd21a9817ee3525980fcb43c1f7fd942085c3b49aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.503009 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.522340 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.535866 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.542983 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.543009 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.543017 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.543032 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.543043 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:05Z","lastTransitionTime":"2025-12-04T03:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.551426 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.563206 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.572526 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.589272 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:56:02Z\\\",\\\"message\\\":\\\"leStack,ClusterIPs:[10.217.4.213],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1204 03:56:02.434417 6668 lb_config.go:1031] Cluster endpoints for openshift-console/downloads for network=default are: map[]\\\\nI1204 03:56:02.434426 6668 services_controller.go:443] Built service openshift-console/downloads LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.213\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:80, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1204 03:56:02.434437 6668 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:56:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.599229 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b9khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41fdb600-3868-48ab-8396-17c2382b6168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b9khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.615060 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.628786 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.639548 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6649e412-5f28-4771-9d38-db928b07e8a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d42708d1f2974b0a6eb579a33c340969fa6df1100355754cb12ceb117f305f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.645172 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.645204 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.645212 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.645242 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.645252 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:05Z","lastTransitionTime":"2025-12-04T03:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.654648 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"186beef2-bf40-41df-ad1d-870b99e5b6f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59760c3aea1ddf4bdfeda36dd7e6e20bb56985e39f27864475a064d0061325de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd1df0baab4bc2ca707440eaa20c66a5b003a783d48ac8ed0c66d185a6e4d116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4e4be31ada4276e640bcf04ab82968c378cfaa920bc82ce54d925d457c51585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.668797 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.681599 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.697432 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:05Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.748203 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.748268 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.748279 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.748296 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.748306 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:05Z","lastTransitionTime":"2025-12-04T03:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.852882 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.852999 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.853015 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.853037 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.853052 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:05Z","lastTransitionTime":"2025-12-04T03:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.955439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.955480 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.955488 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.955503 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:05 crc kubenswrapper[4806]: I1204 03:56:05.955513 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:05Z","lastTransitionTime":"2025-12-04T03:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.057847 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.057895 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.057906 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.057936 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.057946 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:06Z","lastTransitionTime":"2025-12-04T03:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.161042 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.161086 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.161097 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.161114 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.161125 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:06Z","lastTransitionTime":"2025-12-04T03:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.264434 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.264494 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.264505 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.264532 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.264544 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:06Z","lastTransitionTime":"2025-12-04T03:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.368212 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.368280 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.368294 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.368317 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.368334 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:06Z","lastTransitionTime":"2025-12-04T03:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.470811 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.470862 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.470872 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.470892 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.470902 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:06Z","lastTransitionTime":"2025-12-04T03:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.573208 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.573266 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.573277 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.573296 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.573308 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:06Z","lastTransitionTime":"2025-12-04T03:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.675988 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.676061 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.676072 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.676139 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.676154 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:06Z","lastTransitionTime":"2025-12-04T03:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.778282 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.778431 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.778446 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.778465 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.778475 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:06Z","lastTransitionTime":"2025-12-04T03:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.880480 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.880525 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.880538 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.880554 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.880567 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:06Z","lastTransitionTime":"2025-12-04T03:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.983372 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.983416 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.983427 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.983444 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:06 crc kubenswrapper[4806]: I1204 03:56:06.983456 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:06Z","lastTransitionTime":"2025-12-04T03:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.086269 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.086326 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.086339 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.086358 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.086370 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:07Z","lastTransitionTime":"2025-12-04T03:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.188716 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.188753 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.188764 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.188779 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.188789 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:07Z","lastTransitionTime":"2025-12-04T03:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.223368 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.223683 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.223623 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.223946 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.224045 4806 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.223812 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.224172 4806 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.224192 4806 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.224338 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.224151751 +0000 UTC m=+146.082664679 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.224428 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.224418799 +0000 UTC m=+146.082931727 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.291684 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.292054 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.292163 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.292274 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.292358 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:07Z","lastTransitionTime":"2025-12-04T03:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.325666 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.325828 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.325881 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.326052 4806 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.326125 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.326106782 +0000 UTC m=+146.184619710 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.326313 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.326283917 +0000 UTC m=+146.184796845 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.326352 4806 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.326507 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.326497784 +0000 UTC m=+146.185010712 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.394471 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.394515 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.394525 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.394540 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.394551 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:07Z","lastTransitionTime":"2025-12-04T03:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.423161 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.423200 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.423186 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.423308 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.423366 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.423415 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.423654 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:07 crc kubenswrapper[4806]: E1204 03:56:07.423787 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.497245 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.497289 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.497302 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.497319 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.497331 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:07Z","lastTransitionTime":"2025-12-04T03:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.599818 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.599866 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.599877 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.599894 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.599906 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:07Z","lastTransitionTime":"2025-12-04T03:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.703191 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.703224 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.703234 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.703249 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.703261 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:07Z","lastTransitionTime":"2025-12-04T03:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.805464 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.805508 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.805518 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.805533 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.805541 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:07Z","lastTransitionTime":"2025-12-04T03:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.908305 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.908343 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.908356 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.908373 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:07 crc kubenswrapper[4806]: I1204 03:56:07.908383 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:07Z","lastTransitionTime":"2025-12-04T03:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.011030 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.011078 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.011089 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.011109 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.011123 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:08Z","lastTransitionTime":"2025-12-04T03:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.115252 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.115301 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.115310 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.115325 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.115336 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:08Z","lastTransitionTime":"2025-12-04T03:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.217957 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.218138 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.218154 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.218176 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.218190 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:08Z","lastTransitionTime":"2025-12-04T03:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.240292 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.240357 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.240367 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.240387 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.240401 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:08Z","lastTransitionTime":"2025-12-04T03:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:08 crc kubenswrapper[4806]: E1204 03:56:08.256136 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.261199 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.261264 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.261276 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.261299 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.261314 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:08Z","lastTransitionTime":"2025-12-04T03:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:08 crc kubenswrapper[4806]: E1204 03:56:08.275956 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.281167 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.281230 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.281245 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.281262 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.281275 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:08Z","lastTransitionTime":"2025-12-04T03:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:08 crc kubenswrapper[4806]: E1204 03:56:08.296383 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.302441 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.302505 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.302522 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.302547 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.302565 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:08Z","lastTransitionTime":"2025-12-04T03:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:08 crc kubenswrapper[4806]: E1204 03:56:08.318105 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.323779 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.323833 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.323843 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.323863 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.323876 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:08Z","lastTransitionTime":"2025-12-04T03:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:08 crc kubenswrapper[4806]: E1204 03:56:08.338392 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:08Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:08 crc kubenswrapper[4806]: E1204 03:56:08.338570 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.341053 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.341087 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.341106 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.341127 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.341144 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:08Z","lastTransitionTime":"2025-12-04T03:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.443942 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.443985 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.443995 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.444012 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.444024 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:08Z","lastTransitionTime":"2025-12-04T03:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.546580 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.546633 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.546643 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.546662 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.546675 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:08Z","lastTransitionTime":"2025-12-04T03:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.650196 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.650255 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.650271 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.650297 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.650316 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:08Z","lastTransitionTime":"2025-12-04T03:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.756119 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.756173 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.756185 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.756208 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.756223 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:08Z","lastTransitionTime":"2025-12-04T03:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.858955 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.859001 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.859010 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.859024 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.859033 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:08Z","lastTransitionTime":"2025-12-04T03:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.962117 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.962157 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.962167 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.962180 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:08 crc kubenswrapper[4806]: I1204 03:56:08.962190 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:08Z","lastTransitionTime":"2025-12-04T03:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.064337 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.064421 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.064431 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.064445 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.064454 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:09Z","lastTransitionTime":"2025-12-04T03:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.166515 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.166543 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.166551 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.166564 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.166572 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:09Z","lastTransitionTime":"2025-12-04T03:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.268502 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.268542 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.268550 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.268562 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.268582 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:09Z","lastTransitionTime":"2025-12-04T03:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.371231 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.371286 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.371297 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.371318 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.371329 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:09Z","lastTransitionTime":"2025-12-04T03:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.423637 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.423739 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:09 crc kubenswrapper[4806]: E1204 03:56:09.423785 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:09 crc kubenswrapper[4806]: E1204 03:56:09.423884 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.423986 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:09 crc kubenswrapper[4806]: E1204 03:56:09.424036 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.424156 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:09 crc kubenswrapper[4806]: E1204 03:56:09.424365 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.474390 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.474461 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.474497 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.474511 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.474520 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:09Z","lastTransitionTime":"2025-12-04T03:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.577056 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.577110 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.577124 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.577143 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.577155 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:09Z","lastTransitionTime":"2025-12-04T03:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.679755 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.679813 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.679824 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.679841 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.679853 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:09Z","lastTransitionTime":"2025-12-04T03:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.782824 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.782877 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.782888 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.782906 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.782938 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:09Z","lastTransitionTime":"2025-12-04T03:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.885971 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.886020 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.886036 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.886052 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.886063 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:09Z","lastTransitionTime":"2025-12-04T03:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.989007 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.989060 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.989072 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.989093 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:09 crc kubenswrapper[4806]: I1204 03:56:09.989105 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:09Z","lastTransitionTime":"2025-12-04T03:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.092139 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.092274 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.092289 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.092307 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.092319 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:10Z","lastTransitionTime":"2025-12-04T03:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.195442 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.195507 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.195529 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.195573 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.195621 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:10Z","lastTransitionTime":"2025-12-04T03:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.299514 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.299553 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.299563 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.299577 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.299589 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:10Z","lastTransitionTime":"2025-12-04T03:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.403619 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.404116 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.404128 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.404173 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.404184 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:10Z","lastTransitionTime":"2025-12-04T03:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.507090 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.507134 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.507147 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.507162 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.507173 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:10Z","lastTransitionTime":"2025-12-04T03:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.610403 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.610472 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.610486 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.610510 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.610524 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:10Z","lastTransitionTime":"2025-12-04T03:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.712553 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.712607 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.712618 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.712634 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.712656 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:10Z","lastTransitionTime":"2025-12-04T03:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.845958 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.846012 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.846022 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.846041 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.846051 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:10Z","lastTransitionTime":"2025-12-04T03:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.948896 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.948985 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.948997 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.949016 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:10 crc kubenswrapper[4806]: I1204 03:56:10.949046 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:10Z","lastTransitionTime":"2025-12-04T03:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.052669 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.052741 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.052755 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.052783 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.052805 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:11Z","lastTransitionTime":"2025-12-04T03:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.155695 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.155783 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.155806 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.155836 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.155857 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:11Z","lastTransitionTime":"2025-12-04T03:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.260202 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.260339 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.260357 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.260388 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.260406 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:11Z","lastTransitionTime":"2025-12-04T03:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.363376 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.363429 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.363441 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.363463 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.363487 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:11Z","lastTransitionTime":"2025-12-04T03:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.423559 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.423701 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.423796 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:11 crc kubenswrapper[4806]: E1204 03:56:11.423803 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.423857 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:11 crc kubenswrapper[4806]: E1204 03:56:11.424017 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:11 crc kubenswrapper[4806]: E1204 03:56:11.424095 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:11 crc kubenswrapper[4806]: E1204 03:56:11.424163 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.465725 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.465778 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.465789 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.465806 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.465816 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:11Z","lastTransitionTime":"2025-12-04T03:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.568141 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.568193 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.568203 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.568220 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.568231 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:11Z","lastTransitionTime":"2025-12-04T03:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.670721 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.670779 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.670793 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.670811 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.670823 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:11Z","lastTransitionTime":"2025-12-04T03:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.773185 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.773275 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.773290 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.773305 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.773319 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:11Z","lastTransitionTime":"2025-12-04T03:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.876011 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.876059 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.876069 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.876085 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.876098 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:11Z","lastTransitionTime":"2025-12-04T03:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.978252 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.978294 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.978305 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.978321 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:11 crc kubenswrapper[4806]: I1204 03:56:11.978335 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:11Z","lastTransitionTime":"2025-12-04T03:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.081177 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.081232 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.081244 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.081259 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.081269 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:12Z","lastTransitionTime":"2025-12-04T03:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.184415 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.184468 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.184477 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.184493 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.184503 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:12Z","lastTransitionTime":"2025-12-04T03:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.287157 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.287211 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.287223 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.287245 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.287257 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:12Z","lastTransitionTime":"2025-12-04T03:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.389333 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.389381 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.389394 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.389410 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.389424 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:12Z","lastTransitionTime":"2025-12-04T03:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.492036 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.492081 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.492095 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.492113 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.492124 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:12Z","lastTransitionTime":"2025-12-04T03:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.595022 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.595097 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.595109 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.595126 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.595137 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:12Z","lastTransitionTime":"2025-12-04T03:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.698972 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.699032 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.699044 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.699065 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.699077 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:12Z","lastTransitionTime":"2025-12-04T03:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.801809 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.801860 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.801873 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.801890 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.801904 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:12Z","lastTransitionTime":"2025-12-04T03:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.905048 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.905093 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.905115 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.905132 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:12 crc kubenswrapper[4806]: I1204 03:56:12.905143 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:12Z","lastTransitionTime":"2025-12-04T03:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.008239 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.008281 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.008291 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.008306 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.008318 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:13Z","lastTransitionTime":"2025-12-04T03:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.111230 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.111302 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.111315 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.111332 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.111345 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:13Z","lastTransitionTime":"2025-12-04T03:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.213943 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.213995 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.214013 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.214031 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.214041 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:13Z","lastTransitionTime":"2025-12-04T03:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.316842 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.316940 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.316950 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.316963 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.316984 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:13Z","lastTransitionTime":"2025-12-04T03:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.419587 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.419674 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.419690 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.419708 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.419720 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:13Z","lastTransitionTime":"2025-12-04T03:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.422970 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.423036 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.423060 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:13 crc kubenswrapper[4806]: E1204 03:56:13.423094 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.423180 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:13 crc kubenswrapper[4806]: E1204 03:56:13.423280 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:13 crc kubenswrapper[4806]: E1204 03:56:13.423405 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:13 crc kubenswrapper[4806]: E1204 03:56:13.423519 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.522146 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.522185 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.522196 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.522210 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.522223 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:13Z","lastTransitionTime":"2025-12-04T03:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.624807 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.624847 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.624857 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.624875 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.624887 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:13Z","lastTransitionTime":"2025-12-04T03:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.728045 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.728117 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.728141 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.728169 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.728191 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:13Z","lastTransitionTime":"2025-12-04T03:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.831383 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.831434 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.831449 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.831475 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.831499 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:13Z","lastTransitionTime":"2025-12-04T03:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.933814 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.933869 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.933881 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.933909 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:13 crc kubenswrapper[4806]: I1204 03:56:13.933919 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:13Z","lastTransitionTime":"2025-12-04T03:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.036837 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.036915 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.036944 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.036961 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.036976 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:14Z","lastTransitionTime":"2025-12-04T03:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.139596 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.139661 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.139676 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.139698 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.139714 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:14Z","lastTransitionTime":"2025-12-04T03:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.242274 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.243167 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.243227 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.243287 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.243306 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:14Z","lastTransitionTime":"2025-12-04T03:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.346586 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.346627 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.346636 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.346648 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.346658 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:14Z","lastTransitionTime":"2025-12-04T03:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.453469 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.453568 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.453581 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.453599 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.453612 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:14Z","lastTransitionTime":"2025-12-04T03:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.555395 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.555478 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.555498 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.555514 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.555525 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:14Z","lastTransitionTime":"2025-12-04T03:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.658729 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.658768 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.658775 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.658787 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.658796 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:14Z","lastTransitionTime":"2025-12-04T03:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.761681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.761750 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.761775 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.761802 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.761825 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:14Z","lastTransitionTime":"2025-12-04T03:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.864877 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.864966 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.864984 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.865001 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.865014 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:14Z","lastTransitionTime":"2025-12-04T03:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.967740 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.967796 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.967809 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.967826 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:14 crc kubenswrapper[4806]: I1204 03:56:14.967840 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:14Z","lastTransitionTime":"2025-12-04T03:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.070436 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.070475 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.070485 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.070500 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.070511 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:15Z","lastTransitionTime":"2025-12-04T03:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.173026 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.173056 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.173063 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.173075 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.173084 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:15Z","lastTransitionTime":"2025-12-04T03:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.276272 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.276330 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.276347 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.276367 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.276381 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:15Z","lastTransitionTime":"2025-12-04T03:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.379444 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.379497 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.379508 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.379527 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.379537 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:15Z","lastTransitionTime":"2025-12-04T03:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.424135 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.424237 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.424305 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:15 crc kubenswrapper[4806]: E1204 03:56:15.424460 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.424497 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:15 crc kubenswrapper[4806]: E1204 03:56:15.424581 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:15 crc kubenswrapper[4806]: E1204 03:56:15.424756 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:15 crc kubenswrapper[4806]: E1204 03:56:15.424798 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.439423 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c36f919339b78b612aa7c05792af8c9e44d2bb737870d1cb83475eb056be38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.451770 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hqps5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a1fde59-696b-48ad-8fce-491ab2ee8454\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f53772a983569323bfade86c2740c73c36cd87f006e2379eae4a27c6c84494fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-94bvt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hqps5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.475556 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"087077d3-f9fa-4d16-aa7c-eb4241cbb293\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:56:02Z\\\",\\\"message\\\":\\\"leStack,ClusterIPs:[10.217.4.213],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1204 03:56:02.434417 6668 lb_config.go:1031] Cluster endpoints for openshift-console/downloads for network=default are: map[]\\\\nI1204 03:56:02.434426 6668 services_controller.go:443] Built service openshift-console/downloads LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.213\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:80, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1204 03:56:02.434437 6668 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:56:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j8ccw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ljxsg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.482360 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.482391 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.482399 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.482413 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.482423 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:15Z","lastTransitionTime":"2025-12-04T03:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.488514 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b9khd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41fdb600-3868-48ab-8396-17c2382b6168\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xd5dc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:19Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b9khd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.502517 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6839204-3613-4586-bdfc-dbee95dafc68\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8caee0028884135548eb4a826f4cee7cd466adf403cd4c26da82bffb7f78525e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://efa9a5cc87e9db82a8b4f804de0ef5c0cf578d0086a66ddacb4d8ee004c02d10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://29212f06ec4f1b604ced3976a02f1620cabeb2b63b21bf95668ec323f1d48dff\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.516717 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6649e412-5f28-4771-9d38-db928b07e8a2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://63d42708d1f2974b0a6eb579a33c340969fa6df1100355754cb12ceb117f305f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b316ad48a591e7cdb548875951809d1bdaefc5eddc8662f188d3b94e8de3f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.531146 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"186beef2-bf40-41df-ad1d-870b99e5b6f3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59760c3aea1ddf4bdfeda36dd7e6e20bb56985e39f27864475a064d0061325de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd1df0baab4bc2ca707440eaa20c66a5b003a783d48ac8ed0c66d185a6e4d116\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4e4be31ada4276e640bcf04ab82968c378cfaa920bc82ce54d925d457c51585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c98780ab2227edcdc68fb3ca4903f1427442b955b44706a78457c5f00dc9ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.545302 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.563538 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://daa3167171a533a0a8dd473f8b7e4ff81cb04c1924167b49cb96fa1c8fd17848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa7fa75138c28c9781fc102b1856939150d1951732ef1759a322bc89540b00f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.577637 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0f9623b-44ce-43c2-b990-a75db2d17ff8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01965cdecc6cf6e1898b4b55074be1f51b532c7649029bf38683f58a3262edd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmqn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zpps8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.584725 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.585055 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.585187 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.585299 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.585380 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:15Z","lastTransitionTime":"2025-12-04T03:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.592946 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"10aef7a5-56dc-4742-87d3-9c87074eb77a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09c5ee3ce419ae570d166ab3fe281762ff4d3818d36ade4b611b12e5b0c10996\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://02f4dd2765386cc8468cdf647df85430d72ef419215c52f46fcb13af2921fa6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a851f3f4adeb59636679e373a74573524d8f9141c69d623dd00618d2195b61c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc853b87b0c99a4729c542e5dab1f4d69dc271fa609654ca0d749ef49b6ea64f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8747b1aa62f013afb2e05c1c9a8350f543658fa6fb4cf42ac200af597f28bf32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d58e116607f4cc4a6e47f28dbb5dd138dd7293d18ced319a6089fa05f2a70429\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2d6c2fa75c44ff5026e25720504bb112de8428dd1532c8ec6263ee1d4e34d3b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:55:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2j54q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfj6c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.607539 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f45740b4-44fc-497c-9370-ad8b52fe52b4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.620977 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-dsclk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a6e29eb2e4c712cf0fcf618844fe548f0c38982e08169a900c816f7a9a6b71ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T03:55:52Z\\\",\\\"message\\\":\\\"2025-12-04T03:55:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_00ac28ad-f9e0-4f82-9235-011a62af4d92\\\\n2025-12-04T03:55:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_00ac28ad-f9e0-4f82-9235-011a62af4d92 to /host/opt/cni/bin/\\\\n2025-12-04T03:55:06Z [verbose] multus-daemon started\\\\n2025-12-04T03:55:06Z [verbose] Readiness Indicator file check\\\\n2025-12-04T03:55:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T03:55:06Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z858j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-dsclk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.631418 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5d7e1e32-7021-43cb-a282-653d6c4dd750\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2a904eaf2fe5b52cb25f7debdffa162843f726f32575dc4eedd9039a70c1e0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://554a3c8fd3458965e9db603e9b045cb2192b0a226a227822de7075930d727066\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f62m9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:18Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-lgrcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.643606 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.656529 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://562334ce0c67b1e46418c615ade857ff39e91aff36f777235b7de9259e2d09e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.668628 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.678413 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-qdpsj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a576e68-334a-4a4f-aae0-888b6293dc41\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583a860f8fa24b897ae9382ed048bad0eee110b57e85a7bd05c8126e000950c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v229f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:55:04Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-qdpsj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.688554 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.688612 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.688626 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.688646 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.688661 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:15Z","lastTransitionTime":"2025-12-04T03:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.696856 4806 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db305f2-dcdf-4e2c-94ae-0469ce8e27a9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:55:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T03:54:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cc26e5377fcd914affc21e46397fd6ebe234488cc0eadac1c4e9b543301fca2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://87e159737687699b84f24ff0e37415ca1728f856572a87b95e5588537606f0ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f2f51aea4a8f01c7d96bc057fef542163504caf5a97320bf382e60eabaec8cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ec6de5f42f8d12b7fb0770059775b12e82faad5a81e2c0e9ef85f6212cffae7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://25f41c8816799ba262455a943cacc42000e10bcea3a3f90ebca0036e42d5927b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T03:54:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://966a63a12a48455423a3f77e6055886b344e0215cc16c2a2d89bdf1ed967f06d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://966a63a12a48455423a3f77e6055886b344e0215cc16c2a2d89bdf1ed967f06d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e851dfa78c781cb95e3e5a3ea73c38d0c4f1059eebace9a6e17b78a2fc879b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e851dfa78c781cb95e3e5a3ea73c38d0c4f1059eebace9a6e17b78a2fc879b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:46Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://90ad1f66cf96de520eed60dd21a9817ee3525980fcb43c1f7fd942085c3b49aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ad1f66cf96de520eed60dd21a9817ee3525980fcb43c1f7fd942085c3b49aa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T03:54:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T03:54:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T03:54:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:15Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.791754 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.791791 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.791802 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.791820 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.791831 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:15Z","lastTransitionTime":"2025-12-04T03:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.893832 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.893871 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.893882 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.893898 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.893909 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:15Z","lastTransitionTime":"2025-12-04T03:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.996821 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.996877 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.996889 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.996907 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:15 crc kubenswrapper[4806]: I1204 03:56:15.996933 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:15Z","lastTransitionTime":"2025-12-04T03:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.099915 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.099974 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.099987 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.100004 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.100014 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:16Z","lastTransitionTime":"2025-12-04T03:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.202681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.202744 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.202753 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.202772 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.202785 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:16Z","lastTransitionTime":"2025-12-04T03:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.305295 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.305367 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.305384 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.305409 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.305424 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:16Z","lastTransitionTime":"2025-12-04T03:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.408267 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.408321 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.408331 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.408350 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.408366 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:16Z","lastTransitionTime":"2025-12-04T03:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.511344 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.511395 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.511408 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.511426 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.511439 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:16Z","lastTransitionTime":"2025-12-04T03:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.615285 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.615345 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.615357 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.615377 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.615392 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:16Z","lastTransitionTime":"2025-12-04T03:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.717525 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.717581 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.717591 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.717604 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.717612 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:16Z","lastTransitionTime":"2025-12-04T03:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.820265 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.820537 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.820549 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.820564 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.820573 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:16Z","lastTransitionTime":"2025-12-04T03:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.923683 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.923736 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.923747 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.923766 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:16 crc kubenswrapper[4806]: I1204 03:56:16.923783 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:16Z","lastTransitionTime":"2025-12-04T03:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.027392 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.027441 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.027453 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.027472 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.027484 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:17Z","lastTransitionTime":"2025-12-04T03:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.130458 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.130504 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.130519 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.130534 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.130544 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:17Z","lastTransitionTime":"2025-12-04T03:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.233368 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.233437 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.233461 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.233491 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.233512 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:17Z","lastTransitionTime":"2025-12-04T03:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.336341 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.336388 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.336399 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.336415 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.336425 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:17Z","lastTransitionTime":"2025-12-04T03:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.422705 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.422710 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:17 crc kubenswrapper[4806]: E1204 03:56:17.424263 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.422839 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:17 crc kubenswrapper[4806]: E1204 03:56:17.424377 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.422752 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:17 crc kubenswrapper[4806]: E1204 03:56:17.424651 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:17 crc kubenswrapper[4806]: E1204 03:56:17.424715 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.438604 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.438639 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.438648 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.438659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.438668 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:17Z","lastTransitionTime":"2025-12-04T03:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.541671 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.541722 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.541739 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.541756 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.541771 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:17Z","lastTransitionTime":"2025-12-04T03:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.644720 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.644783 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.644796 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.644816 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.644827 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:17Z","lastTransitionTime":"2025-12-04T03:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.747853 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.747899 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.747910 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.747950 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.747963 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:17Z","lastTransitionTime":"2025-12-04T03:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.850995 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.851409 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.851492 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.851564 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.851643 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:17Z","lastTransitionTime":"2025-12-04T03:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.953890 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.954308 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.954640 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.954798 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:17 crc kubenswrapper[4806]: I1204 03:56:17.955054 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:17Z","lastTransitionTime":"2025-12-04T03:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.058120 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.058178 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.058194 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.058213 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.058227 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:18Z","lastTransitionTime":"2025-12-04T03:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.160888 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.161024 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.161061 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.161086 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.161103 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:18Z","lastTransitionTime":"2025-12-04T03:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.264775 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.264839 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.264854 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.264879 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.264898 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:18Z","lastTransitionTime":"2025-12-04T03:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.370224 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.370271 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.370282 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.370301 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.370312 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:18Z","lastTransitionTime":"2025-12-04T03:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.472837 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.472883 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.472894 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.472912 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.472939 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:18Z","lastTransitionTime":"2025-12-04T03:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.492821 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.492865 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.492874 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.492889 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.492900 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:18Z","lastTransitionTime":"2025-12-04T03:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:18 crc kubenswrapper[4806]: E1204 03:56:18.507481 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.512799 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.512854 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.512866 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.512883 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.512896 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:18Z","lastTransitionTime":"2025-12-04T03:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:18 crc kubenswrapper[4806]: E1204 03:56:18.527473 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.531616 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.531650 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.531662 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.531678 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.531690 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:18Z","lastTransitionTime":"2025-12-04T03:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:18 crc kubenswrapper[4806]: E1204 03:56:18.548183 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.552336 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.552393 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.552412 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.552430 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.552441 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:18Z","lastTransitionTime":"2025-12-04T03:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:18 crc kubenswrapper[4806]: E1204 03:56:18.567366 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.572185 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.572236 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.572246 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.572265 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.572277 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:18Z","lastTransitionTime":"2025-12-04T03:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:18 crc kubenswrapper[4806]: E1204 03:56:18.585479 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T03:56:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5dbc708a-80bd-449f-9c89-8b2b0f364021\\\",\\\"systemUUID\\\":\\\"684cf60c-c6df-4023-948c-ef14ad787a60\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T03:56:18Z is after 2025-08-24T17:21:41Z" Dec 04 03:56:18 crc kubenswrapper[4806]: E1204 03:56:18.585705 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.588572 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.588627 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.588639 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.588660 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.588675 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:18Z","lastTransitionTime":"2025-12-04T03:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.691450 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.691512 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.691527 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.691549 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.691564 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:18Z","lastTransitionTime":"2025-12-04T03:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.794948 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.795004 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.795015 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.795031 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.795042 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:18Z","lastTransitionTime":"2025-12-04T03:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.897083 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.897123 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.897134 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.897150 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:18 crc kubenswrapper[4806]: I1204 03:56:18.897162 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:18Z","lastTransitionTime":"2025-12-04T03:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.000755 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.000835 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.000849 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.000866 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.000902 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:19Z","lastTransitionTime":"2025-12-04T03:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.103800 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.103856 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.103873 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.103888 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.103900 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:19Z","lastTransitionTime":"2025-12-04T03:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.206255 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.206308 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.206322 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.206338 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.206350 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:19Z","lastTransitionTime":"2025-12-04T03:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.309172 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.309235 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.309252 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.309274 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.309292 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:19Z","lastTransitionTime":"2025-12-04T03:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.412138 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.412182 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.412191 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.412206 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.412215 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:19Z","lastTransitionTime":"2025-12-04T03:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.425167 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.425268 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.425358 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.425350 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:19 crc kubenswrapper[4806]: E1204 03:56:19.426184 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:19 crc kubenswrapper[4806]: E1204 03:56:19.427026 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:19 crc kubenswrapper[4806]: E1204 03:56:19.427312 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:19 crc kubenswrapper[4806]: E1204 03:56:19.427968 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.428734 4806 scope.go:117] "RemoveContainer" containerID="7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d" Dec 04 03:56:19 crc kubenswrapper[4806]: E1204 03:56:19.429587 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.515339 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.515391 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.515400 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.515417 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.515427 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:19Z","lastTransitionTime":"2025-12-04T03:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.617590 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.617654 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.617666 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.617688 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.617704 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:19Z","lastTransitionTime":"2025-12-04T03:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.721073 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.721116 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.721124 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.721143 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.721153 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:19Z","lastTransitionTime":"2025-12-04T03:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.823430 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.823474 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.823514 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.823550 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.823561 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:19Z","lastTransitionTime":"2025-12-04T03:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.925998 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.926051 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.926064 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.926083 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:19 crc kubenswrapper[4806]: I1204 03:56:19.926096 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:19Z","lastTransitionTime":"2025-12-04T03:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.029094 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.029145 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.029156 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.029190 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.029203 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:20Z","lastTransitionTime":"2025-12-04T03:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.132095 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.132150 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.132159 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.132178 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.132191 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:20Z","lastTransitionTime":"2025-12-04T03:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.234610 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.234682 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.234695 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.234718 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.234732 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:20Z","lastTransitionTime":"2025-12-04T03:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.337465 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.337496 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.337504 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.337516 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.337524 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:20Z","lastTransitionTime":"2025-12-04T03:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.440098 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.440161 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.440173 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.440192 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.440207 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:20Z","lastTransitionTime":"2025-12-04T03:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.543286 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.543334 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.543344 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.543362 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.543373 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:20Z","lastTransitionTime":"2025-12-04T03:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.646191 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.646227 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.646236 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.646251 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.646261 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:20Z","lastTransitionTime":"2025-12-04T03:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.748768 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.748808 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.748817 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.748831 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.748841 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:20Z","lastTransitionTime":"2025-12-04T03:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.851617 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.851663 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.851673 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.851687 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.851697 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:20Z","lastTransitionTime":"2025-12-04T03:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.954518 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.954751 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.954762 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.954776 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:20 crc kubenswrapper[4806]: I1204 03:56:20.954787 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:20Z","lastTransitionTime":"2025-12-04T03:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.058323 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.058410 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.058425 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.058441 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.058451 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:21Z","lastTransitionTime":"2025-12-04T03:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.161247 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.161292 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.161303 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.161318 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.161332 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:21Z","lastTransitionTime":"2025-12-04T03:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.263028 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.263132 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.263144 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.263159 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.263169 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:21Z","lastTransitionTime":"2025-12-04T03:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.365364 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.365399 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.365410 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.365427 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.365438 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:21Z","lastTransitionTime":"2025-12-04T03:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.423741 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.423822 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.423785 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.423741 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:21 crc kubenswrapper[4806]: E1204 03:56:21.423911 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:21 crc kubenswrapper[4806]: E1204 03:56:21.423985 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:21 crc kubenswrapper[4806]: E1204 03:56:21.424139 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:21 crc kubenswrapper[4806]: E1204 03:56:21.424430 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.468875 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.468944 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.468958 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.468976 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.468993 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:21Z","lastTransitionTime":"2025-12-04T03:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.572344 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.572395 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.572406 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.572424 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.572436 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:21Z","lastTransitionTime":"2025-12-04T03:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.675335 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.675408 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.675447 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.675475 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.675492 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:21Z","lastTransitionTime":"2025-12-04T03:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.777749 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.777817 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.777827 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.777841 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.777852 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:21Z","lastTransitionTime":"2025-12-04T03:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.881439 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.881499 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.881511 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.881527 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.881539 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:21Z","lastTransitionTime":"2025-12-04T03:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.985353 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.985522 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.985543 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.985611 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:21 crc kubenswrapper[4806]: I1204 03:56:21.985639 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:21Z","lastTransitionTime":"2025-12-04T03:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.088810 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.088847 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.088857 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.088872 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.088883 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:22Z","lastTransitionTime":"2025-12-04T03:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.191753 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.191801 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.191811 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.191849 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.191864 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:22Z","lastTransitionTime":"2025-12-04T03:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.294615 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.294659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.294670 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.294686 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.294697 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:22Z","lastTransitionTime":"2025-12-04T03:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.397513 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.397565 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.397577 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.397598 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.397611 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:22Z","lastTransitionTime":"2025-12-04T03:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.500601 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.500681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.500697 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.500716 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.500762 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:22Z","lastTransitionTime":"2025-12-04T03:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.603084 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.603128 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.603136 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.603150 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.603159 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:22Z","lastTransitionTime":"2025-12-04T03:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.705643 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.705700 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.705711 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.705726 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.705738 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:22Z","lastTransitionTime":"2025-12-04T03:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.812086 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.812155 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.812168 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.812187 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.812206 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:22Z","lastTransitionTime":"2025-12-04T03:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.915485 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.915524 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.915534 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.915548 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:22 crc kubenswrapper[4806]: I1204 03:56:22.915559 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:22Z","lastTransitionTime":"2025-12-04T03:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.018807 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.018869 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.018881 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.018899 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.018912 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:23Z","lastTransitionTime":"2025-12-04T03:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.122320 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.122366 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.122377 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.122393 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.122406 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:23Z","lastTransitionTime":"2025-12-04T03:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.226294 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.226357 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.226371 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.226393 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.226407 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:23Z","lastTransitionTime":"2025-12-04T03:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.329749 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.329835 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.329858 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.329876 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.329890 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:23Z","lastTransitionTime":"2025-12-04T03:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.423714 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.423786 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.423844 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:23 crc kubenswrapper[4806]: E1204 03:56:23.424058 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.424109 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:23 crc kubenswrapper[4806]: E1204 03:56:23.424376 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:23 crc kubenswrapper[4806]: E1204 03:56:23.424431 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:23 crc kubenswrapper[4806]: E1204 03:56:23.424565 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.432444 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.432492 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.432504 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.432522 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.432536 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:23Z","lastTransitionTime":"2025-12-04T03:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.536018 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.536071 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.536083 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.536103 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.536116 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:23Z","lastTransitionTime":"2025-12-04T03:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.639575 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.639644 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.639658 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.639681 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.639695 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:23Z","lastTransitionTime":"2025-12-04T03:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.703501 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs\") pod \"network-metrics-daemon-b9khd\" (UID: \"41fdb600-3868-48ab-8396-17c2382b6168\") " pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:23 crc kubenswrapper[4806]: E1204 03:56:23.703729 4806 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 03:56:23 crc kubenswrapper[4806]: E1204 03:56:23.703843 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs podName:41fdb600-3868-48ab-8396-17c2382b6168 nodeName:}" failed. No retries permitted until 2025-12-04 03:57:27.703821254 +0000 UTC m=+162.562334242 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs") pod "network-metrics-daemon-b9khd" (UID: "41fdb600-3868-48ab-8396-17c2382b6168") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.742719 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.742776 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.742788 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.742805 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.742817 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:23Z","lastTransitionTime":"2025-12-04T03:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.846073 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.846132 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.846144 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.846166 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.846181 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:23Z","lastTransitionTime":"2025-12-04T03:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.949503 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.949667 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.949680 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.949697 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:23 crc kubenswrapper[4806]: I1204 03:56:23.949707 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:23Z","lastTransitionTime":"2025-12-04T03:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.053032 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.053094 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.053105 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.053126 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.053145 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:24Z","lastTransitionTime":"2025-12-04T03:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.155361 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.155412 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.155426 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.155442 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.155454 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:24Z","lastTransitionTime":"2025-12-04T03:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.259226 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.259270 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.259281 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.259297 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.259308 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:24Z","lastTransitionTime":"2025-12-04T03:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.362034 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.362112 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.362130 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.362153 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.362168 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:24Z","lastTransitionTime":"2025-12-04T03:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.465968 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.466030 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.466049 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.466072 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.466086 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:24Z","lastTransitionTime":"2025-12-04T03:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.569532 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.569591 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.569605 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.569629 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.569644 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:24Z","lastTransitionTime":"2025-12-04T03:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.672690 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.672732 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.672743 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.672759 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.672771 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:24Z","lastTransitionTime":"2025-12-04T03:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.775392 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.775449 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.775460 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.775478 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.775489 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:24Z","lastTransitionTime":"2025-12-04T03:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.877697 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.877770 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.877779 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.877793 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.877802 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:24Z","lastTransitionTime":"2025-12-04T03:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.980231 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.980283 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.980294 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.980308 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:24 crc kubenswrapper[4806]: I1204 03:56:24.980319 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:24Z","lastTransitionTime":"2025-12-04T03:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.082884 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.082966 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.082976 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.082992 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.083004 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:25Z","lastTransitionTime":"2025-12-04T03:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.186975 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.187021 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.187035 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.187156 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.187172 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:25Z","lastTransitionTime":"2025-12-04T03:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.290130 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.290218 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.290231 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.290277 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.290292 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:25Z","lastTransitionTime":"2025-12-04T03:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.393164 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.393524 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.393536 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.393552 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.393562 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:25Z","lastTransitionTime":"2025-12-04T03:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.423495 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:25 crc kubenswrapper[4806]: E1204 03:56:25.423645 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.424192 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.424461 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:25 crc kubenswrapper[4806]: E1204 03:56:25.424669 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:25 crc kubenswrapper[4806]: E1204 03:56:25.425127 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.425725 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:25 crc kubenswrapper[4806]: E1204 03:56:25.425910 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.462037 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=22.462020053 podStartE2EDuration="22.462020053s" podCreationTimestamp="2025-12-04 03:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:56:25.461464286 +0000 UTC m=+100.319977224" watchObservedRunningTime="2025-12-04 03:56:25.462020053 +0000 UTC m=+100.320532981" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.496261 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.496305 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.496319 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.496339 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.496354 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:25Z","lastTransitionTime":"2025-12-04T03:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.504316 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-qdpsj" podStartSLOduration=81.504296073 podStartE2EDuration="1m21.504296073s" podCreationTimestamp="2025-12-04 03:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:56:25.503812598 +0000 UTC m=+100.362325526" watchObservedRunningTime="2025-12-04 03:56:25.504296073 +0000 UTC m=+100.362809001" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.534047 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=79.534027701 podStartE2EDuration="1m19.534027701s" podCreationTimestamp="2025-12-04 03:55:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:56:25.521962982 +0000 UTC m=+100.380475910" watchObservedRunningTime="2025-12-04 03:56:25.534027701 +0000 UTC m=+100.392540629" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.548168 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-hqps5" podStartSLOduration=81.54814581 podStartE2EDuration="1m21.54814581s" podCreationTimestamp="2025-12-04 03:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:56:25.544546031 +0000 UTC m=+100.403058959" watchObservedRunningTime="2025-12-04 03:56:25.54814581 +0000 UTC m=+100.406658738" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.598844 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.598885 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.598897 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.598913 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.598947 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:25Z","lastTransitionTime":"2025-12-04T03:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.612622 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=82.612598647 podStartE2EDuration="1m22.612598647s" podCreationTimestamp="2025-12-04 03:55:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:56:25.601453347 +0000 UTC m=+100.459966275" watchObservedRunningTime="2025-12-04 03:56:25.612598647 +0000 UTC m=+100.471111575" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.613326 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=33.613319939 podStartE2EDuration="33.613319939s" podCreationTimestamp="2025-12-04 03:55:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:56:25.612902257 +0000 UTC m=+100.471415185" watchObservedRunningTime="2025-12-04 03:56:25.613319939 +0000 UTC m=+100.471832857" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.641910 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=46.641870851 podStartE2EDuration="46.641870851s" podCreationTimestamp="2025-12-04 03:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:56:25.629668908 +0000 UTC m=+100.488181836" watchObservedRunningTime="2025-12-04 03:56:25.641870851 +0000 UTC m=+100.500383779" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.670683 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podStartSLOduration=80.67065788 podStartE2EDuration="1m20.67065788s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:56:25.670520975 +0000 UTC m=+100.529033923" watchObservedRunningTime="2025-12-04 03:56:25.67065788 +0000 UTC m=+100.529170808" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.691782 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-mfj6c" podStartSLOduration=80.691762003 podStartE2EDuration="1m20.691762003s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:56:25.691497195 +0000 UTC m=+100.550010123" watchObservedRunningTime="2025-12-04 03:56:25.691762003 +0000 UTC m=+100.550274921" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.701578 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.701647 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.701656 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.701670 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.701680 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:25Z","lastTransitionTime":"2025-12-04T03:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.723661 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-dsclk" podStartSLOduration=80.723635096 podStartE2EDuration="1m20.723635096s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:56:25.722195591 +0000 UTC m=+100.580708529" watchObservedRunningTime="2025-12-04 03:56:25.723635096 +0000 UTC m=+100.582148024" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.735879 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-lgrcd" podStartSLOduration=80.735856808 podStartE2EDuration="1m20.735856808s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:56:25.734510458 +0000 UTC m=+100.593023386" watchObservedRunningTime="2025-12-04 03:56:25.735856808 +0000 UTC m=+100.594369736" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.804273 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.804322 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.804332 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.804345 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.804373 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:25Z","lastTransitionTime":"2025-12-04T03:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.911291 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.911356 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.911374 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.911397 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:25 crc kubenswrapper[4806]: I1204 03:56:25.911415 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:25Z","lastTransitionTime":"2025-12-04T03:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.014634 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.014673 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.014685 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.014701 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.014713 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:26Z","lastTransitionTime":"2025-12-04T03:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.117966 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.118008 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.118020 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.118038 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.118048 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:26Z","lastTransitionTime":"2025-12-04T03:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.222002 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.222055 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.222064 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.222081 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.222095 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:26Z","lastTransitionTime":"2025-12-04T03:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.324539 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.324600 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.324614 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.324652 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.324666 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:26Z","lastTransitionTime":"2025-12-04T03:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.427042 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.427102 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.427114 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.427133 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.427145 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:26Z","lastTransitionTime":"2025-12-04T03:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.531124 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.531189 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.531202 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.531224 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.531238 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:26Z","lastTransitionTime":"2025-12-04T03:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.633738 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.633793 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.633806 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.633823 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.633853 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:26Z","lastTransitionTime":"2025-12-04T03:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.736771 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.736814 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.736823 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.736839 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.736849 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:26Z","lastTransitionTime":"2025-12-04T03:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.839852 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.839898 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.839908 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.839940 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.839953 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:26Z","lastTransitionTime":"2025-12-04T03:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.942583 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.942619 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.942628 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.942641 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:26 crc kubenswrapper[4806]: I1204 03:56:26.942651 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:26Z","lastTransitionTime":"2025-12-04T03:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.045753 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.045794 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.045804 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.045821 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.045832 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:27Z","lastTransitionTime":"2025-12-04T03:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.148291 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.148343 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.148352 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.148368 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.148378 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:27Z","lastTransitionTime":"2025-12-04T03:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.251005 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.251091 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.251117 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.251147 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.251170 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:27Z","lastTransitionTime":"2025-12-04T03:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.354041 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.354071 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.354079 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.354092 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.354101 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:27Z","lastTransitionTime":"2025-12-04T03:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.423106 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.423426 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.423515 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:27 crc kubenswrapper[4806]: E1204 03:56:27.423670 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.423831 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:27 crc kubenswrapper[4806]: E1204 03:56:27.423910 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:27 crc kubenswrapper[4806]: E1204 03:56:27.423998 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:27 crc kubenswrapper[4806]: E1204 03:56:27.424054 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.457239 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.457283 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.457323 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.457340 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.457353 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:27Z","lastTransitionTime":"2025-12-04T03:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.559015 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.559057 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.559068 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.559084 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.559096 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:27Z","lastTransitionTime":"2025-12-04T03:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.661846 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.661910 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.661966 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.661997 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.662021 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:27Z","lastTransitionTime":"2025-12-04T03:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.764093 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.764145 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.764159 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.764174 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.764187 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:27Z","lastTransitionTime":"2025-12-04T03:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.866902 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.866989 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.867004 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.867026 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.867054 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:27Z","lastTransitionTime":"2025-12-04T03:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.977082 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.977151 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.977164 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.977184 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:27 crc kubenswrapper[4806]: I1204 03:56:27.977197 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:27Z","lastTransitionTime":"2025-12-04T03:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.079696 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.079762 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.079775 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.079799 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.079817 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:28Z","lastTransitionTime":"2025-12-04T03:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.182706 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.182764 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.182775 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.182792 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.182804 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:28Z","lastTransitionTime":"2025-12-04T03:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.285694 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.285840 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.285855 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.285872 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.285884 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:28Z","lastTransitionTime":"2025-12-04T03:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.388659 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.388715 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.388728 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.388747 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.388760 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:28Z","lastTransitionTime":"2025-12-04T03:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.491771 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.491807 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.491817 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.491831 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.491840 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:28Z","lastTransitionTime":"2025-12-04T03:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.594585 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.594648 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.594662 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.594699 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.594713 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:28Z","lastTransitionTime":"2025-12-04T03:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.696778 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.696822 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.696834 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.696851 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.696863 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:28Z","lastTransitionTime":"2025-12-04T03:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.730112 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.730163 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.730175 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.730191 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.730203 4806 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T03:56:28Z","lastTransitionTime":"2025-12-04T03:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.779186 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d"] Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.779576 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.783023 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.783707 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.784155 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.784673 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.850525 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hqp6d\" (UID: \"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.850634 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hqp6d\" (UID: \"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.850685 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hqp6d\" (UID: \"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.850719 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hqp6d\" (UID: \"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.850747 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hqp6d\" (UID: \"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.951695 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hqp6d\" (UID: \"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.951784 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hqp6d\" (UID: \"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.951846 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hqp6d\" (UID: \"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.951912 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hqp6d\" (UID: \"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.952062 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hqp6d\" (UID: \"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.951847 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-hqp6d\" (UID: \"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.951906 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-hqp6d\" (UID: \"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.953532 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc-service-ca\") pod \"cluster-version-operator-5c965bbfc6-hqp6d\" (UID: \"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.968654 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-hqp6d\" (UID: \"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:28 crc kubenswrapper[4806]: I1204 03:56:28.981550 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-hqp6d\" (UID: \"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:29 crc kubenswrapper[4806]: I1204 03:56:29.094647 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" Dec 04 03:56:29 crc kubenswrapper[4806]: I1204 03:56:29.423686 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:29 crc kubenswrapper[4806]: E1204 03:56:29.423852 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:29 crc kubenswrapper[4806]: I1204 03:56:29.423983 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:29 crc kubenswrapper[4806]: I1204 03:56:29.424104 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:29 crc kubenswrapper[4806]: E1204 03:56:29.424106 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:29 crc kubenswrapper[4806]: I1204 03:56:29.424148 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:29 crc kubenswrapper[4806]: E1204 03:56:29.424210 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:29 crc kubenswrapper[4806]: E1204 03:56:29.424274 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:29 crc kubenswrapper[4806]: I1204 03:56:29.924613 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" event={"ID":"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc","Type":"ContainerStarted","Data":"4a7b865a93c201efd4dfb2a835af4474bf56e1fd581e90707b89e957f57c9819"} Dec 04 03:56:29 crc kubenswrapper[4806]: I1204 03:56:29.924679 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" event={"ID":"79e7fdb5-a9f9-4533-aff8-aee8b4faf9fc","Type":"ContainerStarted","Data":"a974dcf7ebd17b992e3b895754af921e9a50b0916c2a1fe67b59ccc53c3fe74e"} Dec 04 03:56:29 crc kubenswrapper[4806]: I1204 03:56:29.939539 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-hqp6d" podStartSLOduration=84.939516817 podStartE2EDuration="1m24.939516817s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:56:29.938884918 +0000 UTC m=+104.797397846" watchObservedRunningTime="2025-12-04 03:56:29.939516817 +0000 UTC m=+104.798029745" Dec 04 03:56:31 crc kubenswrapper[4806]: I1204 03:56:31.423380 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:31 crc kubenswrapper[4806]: I1204 03:56:31.423457 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:31 crc kubenswrapper[4806]: I1204 03:56:31.423530 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:31 crc kubenswrapper[4806]: I1204 03:56:31.423550 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:31 crc kubenswrapper[4806]: E1204 03:56:31.423703 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:31 crc kubenswrapper[4806]: E1204 03:56:31.423839 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:31 crc kubenswrapper[4806]: E1204 03:56:31.424013 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:31 crc kubenswrapper[4806]: E1204 03:56:31.424158 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:32 crc kubenswrapper[4806]: I1204 03:56:32.423773 4806 scope.go:117] "RemoveContainer" containerID="7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d" Dec 04 03:56:32 crc kubenswrapper[4806]: E1204 03:56:32.423953 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ljxsg_openshift-ovn-kubernetes(087077d3-f9fa-4d16-aa7c-eb4241cbb293)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" Dec 04 03:56:33 crc kubenswrapper[4806]: I1204 03:56:33.422816 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:33 crc kubenswrapper[4806]: I1204 03:56:33.423040 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:33 crc kubenswrapper[4806]: E1204 03:56:33.423110 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:33 crc kubenswrapper[4806]: I1204 03:56:33.423187 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:33 crc kubenswrapper[4806]: E1204 03:56:33.423435 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:33 crc kubenswrapper[4806]: I1204 03:56:33.423530 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:33 crc kubenswrapper[4806]: E1204 03:56:33.423548 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:33 crc kubenswrapper[4806]: E1204 03:56:33.423781 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:35 crc kubenswrapper[4806]: I1204 03:56:35.423182 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:35 crc kubenswrapper[4806]: I1204 03:56:35.423242 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:35 crc kubenswrapper[4806]: I1204 03:56:35.423274 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:35 crc kubenswrapper[4806]: E1204 03:56:35.424076 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:35 crc kubenswrapper[4806]: I1204 03:56:35.424099 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:35 crc kubenswrapper[4806]: E1204 03:56:35.424382 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:35 crc kubenswrapper[4806]: E1204 03:56:35.424454 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:35 crc kubenswrapper[4806]: E1204 03:56:35.424407 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:37 crc kubenswrapper[4806]: I1204 03:56:37.423418 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:37 crc kubenswrapper[4806]: E1204 03:56:37.423881 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:37 crc kubenswrapper[4806]: I1204 03:56:37.423475 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:37 crc kubenswrapper[4806]: E1204 03:56:37.424003 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:37 crc kubenswrapper[4806]: I1204 03:56:37.423510 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:37 crc kubenswrapper[4806]: I1204 03:56:37.423432 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:37 crc kubenswrapper[4806]: E1204 03:56:37.424084 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:37 crc kubenswrapper[4806]: E1204 03:56:37.424153 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:38 crc kubenswrapper[4806]: I1204 03:56:38.955599 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dsclk_4f13bcfc-c287-40fd-b8f6-e4bbea8c6577/kube-multus/1.log" Dec 04 03:56:38 crc kubenswrapper[4806]: I1204 03:56:38.956603 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dsclk_4f13bcfc-c287-40fd-b8f6-e4bbea8c6577/kube-multus/0.log" Dec 04 03:56:38 crc kubenswrapper[4806]: I1204 03:56:38.956695 4806 generic.go:334] "Generic (PLEG): container finished" podID="4f13bcfc-c287-40fd-b8f6-e4bbea8c6577" containerID="a6e29eb2e4c712cf0fcf618844fe548f0c38982e08169a900c816f7a9a6b71ba" exitCode=1 Dec 04 03:56:38 crc kubenswrapper[4806]: I1204 03:56:38.956812 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dsclk" event={"ID":"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577","Type":"ContainerDied","Data":"a6e29eb2e4c712cf0fcf618844fe548f0c38982e08169a900c816f7a9a6b71ba"} Dec 04 03:56:38 crc kubenswrapper[4806]: I1204 03:56:38.956895 4806 scope.go:117] "RemoveContainer" containerID="49bb0afd2ed458a48f2882246ca10c99adb3197a527f74c8c96013810bfd923b" Dec 04 03:56:38 crc kubenswrapper[4806]: I1204 03:56:38.958013 4806 scope.go:117] "RemoveContainer" containerID="a6e29eb2e4c712cf0fcf618844fe548f0c38982e08169a900c816f7a9a6b71ba" Dec 04 03:56:38 crc kubenswrapper[4806]: E1204 03:56:38.958468 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-dsclk_openshift-multus(4f13bcfc-c287-40fd-b8f6-e4bbea8c6577)\"" pod="openshift-multus/multus-dsclk" podUID="4f13bcfc-c287-40fd-b8f6-e4bbea8c6577" Dec 04 03:56:39 crc kubenswrapper[4806]: I1204 03:56:39.424043 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:39 crc kubenswrapper[4806]: I1204 03:56:39.424119 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:39 crc kubenswrapper[4806]: I1204 03:56:39.424063 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:39 crc kubenswrapper[4806]: E1204 03:56:39.424266 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:39 crc kubenswrapper[4806]: I1204 03:56:39.424373 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:39 crc kubenswrapper[4806]: E1204 03:56:39.424497 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:39 crc kubenswrapper[4806]: E1204 03:56:39.424610 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:39 crc kubenswrapper[4806]: E1204 03:56:39.424677 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:39 crc kubenswrapper[4806]: I1204 03:56:39.961710 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dsclk_4f13bcfc-c287-40fd-b8f6-e4bbea8c6577/kube-multus/1.log" Dec 04 03:56:41 crc kubenswrapper[4806]: I1204 03:56:41.422632 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:41 crc kubenswrapper[4806]: I1204 03:56:41.422764 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:41 crc kubenswrapper[4806]: E1204 03:56:41.422820 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:41 crc kubenswrapper[4806]: I1204 03:56:41.422964 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:41 crc kubenswrapper[4806]: E1204 03:56:41.423048 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:41 crc kubenswrapper[4806]: I1204 03:56:41.423210 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:41 crc kubenswrapper[4806]: E1204 03:56:41.423210 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:41 crc kubenswrapper[4806]: E1204 03:56:41.423298 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:43 crc kubenswrapper[4806]: I1204 03:56:43.424168 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:43 crc kubenswrapper[4806]: I1204 03:56:43.424202 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:43 crc kubenswrapper[4806]: E1204 03:56:43.424249 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:43 crc kubenswrapper[4806]: I1204 03:56:43.424253 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:43 crc kubenswrapper[4806]: E1204 03:56:43.424384 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:43 crc kubenswrapper[4806]: I1204 03:56:43.424437 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:43 crc kubenswrapper[4806]: E1204 03:56:43.424483 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:43 crc kubenswrapper[4806]: E1204 03:56:43.424607 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:45 crc kubenswrapper[4806]: E1204 03:56:45.380673 4806 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 04 03:56:45 crc kubenswrapper[4806]: I1204 03:56:45.424124 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:45 crc kubenswrapper[4806]: E1204 03:56:45.429470 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:45 crc kubenswrapper[4806]: I1204 03:56:45.429985 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:45 crc kubenswrapper[4806]: I1204 03:56:45.430074 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:45 crc kubenswrapper[4806]: E1204 03:56:45.430166 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:45 crc kubenswrapper[4806]: I1204 03:56:45.430840 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:45 crc kubenswrapper[4806]: E1204 03:56:45.430973 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:45 crc kubenswrapper[4806]: E1204 03:56:45.431410 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:45 crc kubenswrapper[4806]: E1204 03:56:45.520891 4806 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 03:56:46 crc kubenswrapper[4806]: I1204 03:56:46.424665 4806 scope.go:117] "RemoveContainer" containerID="7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d" Dec 04 03:56:46 crc kubenswrapper[4806]: I1204 03:56:46.986813 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/3.log" Dec 04 03:56:46 crc kubenswrapper[4806]: I1204 03:56:46.990634 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerStarted","Data":"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8"} Dec 04 03:56:46 crc kubenswrapper[4806]: I1204 03:56:46.992149 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:56:47 crc kubenswrapper[4806]: I1204 03:56:47.025109 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podStartSLOduration=102.025087068 podStartE2EDuration="1m42.025087068s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:56:47.024235062 +0000 UTC m=+121.882747990" watchObservedRunningTime="2025-12-04 03:56:47.025087068 +0000 UTC m=+121.883599996" Dec 04 03:56:47 crc kubenswrapper[4806]: I1204 03:56:47.269353 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-b9khd"] Dec 04 03:56:47 crc kubenswrapper[4806]: I1204 03:56:47.269524 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:47 crc kubenswrapper[4806]: E1204 03:56:47.269651 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:47 crc kubenswrapper[4806]: I1204 03:56:47.425361 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:47 crc kubenswrapper[4806]: I1204 03:56:47.425521 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:47 crc kubenswrapper[4806]: E1204 03:56:47.425697 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:47 crc kubenswrapper[4806]: I1204 03:56:47.425976 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:47 crc kubenswrapper[4806]: E1204 03:56:47.426210 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:47 crc kubenswrapper[4806]: E1204 03:56:47.426402 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:49 crc kubenswrapper[4806]: I1204 03:56:49.423136 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:49 crc kubenswrapper[4806]: I1204 03:56:49.423147 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:49 crc kubenswrapper[4806]: E1204 03:56:49.423839 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:49 crc kubenswrapper[4806]: I1204 03:56:49.423321 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:49 crc kubenswrapper[4806]: I1204 03:56:49.423169 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:49 crc kubenswrapper[4806]: E1204 03:56:49.424044 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:49 crc kubenswrapper[4806]: E1204 03:56:49.424123 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:49 crc kubenswrapper[4806]: E1204 03:56:49.424260 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:50 crc kubenswrapper[4806]: E1204 03:56:50.522062 4806 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 03:56:51 crc kubenswrapper[4806]: I1204 03:56:51.423141 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:51 crc kubenswrapper[4806]: I1204 03:56:51.423211 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:51 crc kubenswrapper[4806]: E1204 03:56:51.423277 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:51 crc kubenswrapper[4806]: I1204 03:56:51.423349 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:51 crc kubenswrapper[4806]: I1204 03:56:51.423347 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:51 crc kubenswrapper[4806]: E1204 03:56:51.423420 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:51 crc kubenswrapper[4806]: E1204 03:56:51.423499 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:51 crc kubenswrapper[4806]: E1204 03:56:51.423551 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:53 crc kubenswrapper[4806]: I1204 03:56:53.422710 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:53 crc kubenswrapper[4806]: E1204 03:56:53.422871 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:53 crc kubenswrapper[4806]: I1204 03:56:53.422730 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:53 crc kubenswrapper[4806]: I1204 03:56:53.422710 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:53 crc kubenswrapper[4806]: I1204 03:56:53.423267 4806 scope.go:117] "RemoveContainer" containerID="a6e29eb2e4c712cf0fcf618844fe548f0c38982e08169a900c816f7a9a6b71ba" Dec 04 03:56:53 crc kubenswrapper[4806]: E1204 03:56:53.423429 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:53 crc kubenswrapper[4806]: I1204 03:56:53.423574 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:53 crc kubenswrapper[4806]: E1204 03:56:53.423635 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:53 crc kubenswrapper[4806]: E1204 03:56:53.423756 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:54 crc kubenswrapper[4806]: I1204 03:56:54.013068 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dsclk_4f13bcfc-c287-40fd-b8f6-e4bbea8c6577/kube-multus/1.log" Dec 04 03:56:54 crc kubenswrapper[4806]: I1204 03:56:54.013441 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dsclk" event={"ID":"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577","Type":"ContainerStarted","Data":"d6efe030cffcc9ae048a86abeffad5784f713909803cbabe74db70e9e5212015"} Dec 04 03:56:55 crc kubenswrapper[4806]: I1204 03:56:55.423693 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:55 crc kubenswrapper[4806]: I1204 03:56:55.423736 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:55 crc kubenswrapper[4806]: E1204 03:56:55.425662 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 03:56:55 crc kubenswrapper[4806]: I1204 03:56:55.423879 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:55 crc kubenswrapper[4806]: I1204 03:56:55.423779 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:55 crc kubenswrapper[4806]: E1204 03:56:55.425749 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 03:56:55 crc kubenswrapper[4806]: E1204 03:56:55.425939 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b9khd" podUID="41fdb600-3868-48ab-8396-17c2382b6168" Dec 04 03:56:55 crc kubenswrapper[4806]: E1204 03:56:55.425905 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 03:56:55 crc kubenswrapper[4806]: I1204 03:56:55.652290 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 03:56:57 crc kubenswrapper[4806]: I1204 03:56:57.423407 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:56:57 crc kubenswrapper[4806]: I1204 03:56:57.423586 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:56:57 crc kubenswrapper[4806]: I1204 03:56:57.423605 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:56:57 crc kubenswrapper[4806]: I1204 03:56:57.423605 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:56:57 crc kubenswrapper[4806]: I1204 03:56:57.431082 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 04 03:56:57 crc kubenswrapper[4806]: I1204 03:56:57.431343 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 04 03:56:57 crc kubenswrapper[4806]: I1204 03:56:57.431796 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 04 03:56:57 crc kubenswrapper[4806]: I1204 03:56:57.432023 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 04 03:56:57 crc kubenswrapper[4806]: I1204 03:56:57.432170 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 04 03:56:57 crc kubenswrapper[4806]: I1204 03:56:57.432287 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.630621 4806 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.665250 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-jhptt"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.665744 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.667309 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.667726 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.668554 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-62l7r"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.668832 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-62l7r" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.669433 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j8m72"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.669805 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.670561 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-7lbq7"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.671153 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.675563 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8nlkb"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.676160 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8nlkb" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.681894 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.682661 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.684581 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.685547 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.685892 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.686066 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.686849 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.687277 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.687436 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.687468 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.687710 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.687758 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.687797 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.687904 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.687940 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.688008 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.688133 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.688282 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.688408 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.688963 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pddk8"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.689166 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.690358 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.689528 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.689679 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.689734 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.702595 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.708222 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.708544 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.709039 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.709217 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.709106 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.709517 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.709646 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.710539 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.710608 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.710736 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.710811 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.710882 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.711070 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.711224 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.711434 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.711586 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.711702 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.711740 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.711791 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.711846 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.711951 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.711997 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.712060 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.712106 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.712127 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.712164 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.712066 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.712266 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.712298 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.712270 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.712404 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.714024 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7m42k"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.714491 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.715346 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.717594 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.722472 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6qzsg"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.722878 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-8r5hx"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.723120 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c2d5j"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.723406 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.723693 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.723869 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8r5hx" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.726041 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.727619 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.728490 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.728612 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.728667 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.728633 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.728758 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.728843 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.728858 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.728995 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.729039 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.729218 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.729410 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.733531 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.735861 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.736432 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fd998"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.736761 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.737061 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.737084 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.738300 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.739428 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.740580 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.741026 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.741177 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.741801 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-mmxrh"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.742591 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.743152 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.754846 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.742683 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.757994 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.760316 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.760959 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ln6xb"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.761521 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.762388 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.762504 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.762778 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.762995 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.763219 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.765532 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.769435 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.770754 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.770896 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.771108 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ln6xb" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.772131 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.772255 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.772354 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.772418 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.772495 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.772522 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.772612 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.772803 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.772805 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.780037 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.785296 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.787168 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.788166 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.788634 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.789606 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.789837 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.790068 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.790148 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.792391 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.793824 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.796296 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.798127 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.798839 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dgkrj"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.799220 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.799276 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.799763 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-jhptt"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.799869 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.800225 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.800432 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dgkrj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.802408 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.805641 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807511 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b4637a3-8d20-4427-8e71-9b5ab8e6ee8c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8nlkb\" (UID: \"5b4637a3-8d20-4427-8e71-9b5ab8e6ee8c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8nlkb" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807551 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5761a71-91cf-4781-9d7d-dc3f9b9096d0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-z4k8p\" (UID: \"b5761a71-91cf-4781-9d7d-dc3f9b9096d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807579 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6-config\") pod \"machine-api-operator-5694c8668f-j8m72\" (UID: \"7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807600 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w9cv\" (UniqueName: \"kubernetes.io/projected/7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6-kube-api-access-7w9cv\") pod \"machine-api-operator-5694c8668f-j8m72\" (UID: \"7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807623 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5229f915-0d18-41e7-abaf-1d2a783670eb-machine-approver-tls\") pod \"machine-approver-56656f9798-75rsr\" (UID: \"5229f915-0d18-41e7-abaf-1d2a783670eb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807658 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-946xt\" (UniqueName: \"kubernetes.io/projected/b5761a71-91cf-4781-9d7d-dc3f9b9096d0-kube-api-access-946xt\") pod \"openshift-apiserver-operator-796bbdcf4f-z4k8p\" (UID: \"b5761a71-91cf-4781-9d7d-dc3f9b9096d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807681 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/53e674b8-9044-45f1-8aa5-bb81ec63ae4a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-jhptt\" (UID: \"53e674b8-9044-45f1-8aa5-bb81ec63ae4a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807705 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj52k\" (UniqueName: \"kubernetes.io/projected/5b4637a3-8d20-4427-8e71-9b5ab8e6ee8c-kube-api-access-gj52k\") pod \"cluster-samples-operator-665b6dd947-8nlkb\" (UID: \"5b4637a3-8d20-4427-8e71-9b5ab8e6ee8c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8nlkb" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807733 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc5dm\" (UniqueName: \"kubernetes.io/projected/53e674b8-9044-45f1-8aa5-bb81ec63ae4a-kube-api-access-nc5dm\") pod \"openshift-config-operator-7777fb866f-jhptt\" (UID: \"53e674b8-9044-45f1-8aa5-bb81ec63ae4a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807756 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s46kq\" (UniqueName: \"kubernetes.io/projected/606e54dc-5cc4-4e02-b796-dd37b613a8d0-kube-api-access-s46kq\") pod \"controller-manager-879f6c89f-pddk8\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807777 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba31bbad-4048-4651-92cd-d0cba1179541-serving-cert\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807800 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba31bbad-4048-4651-92cd-d0cba1179541-config\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807819 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ba31bbad-4048-4651-92cd-d0cba1179541-etcd-client\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807843 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ca5dd6-f353-4c80-ad4b-1dacf3057d1d-config\") pod \"console-operator-58897d9998-62l7r\" (UID: \"32ca5dd6-f353-4c80-ad4b-1dacf3057d1d\") " pod="openshift-console-operator/console-operator-58897d9998-62l7r" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807864 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-pddk8\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807886 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-client-ca\") pod \"route-controller-manager-6576b87f9c-f68ml\" (UID: \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.807907 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/40553430-c0eb-45b4-880b-d0e1f85adb33-audit-dir\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808016 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zj6n\" (UniqueName: \"kubernetes.io/projected/5229f915-0d18-41e7-abaf-1d2a783670eb-kube-api-access-9zj6n\") pod \"machine-approver-56656f9798-75rsr\" (UID: \"5229f915-0d18-41e7-abaf-1d2a783670eb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808047 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/32ca5dd6-f353-4c80-ad4b-1dacf3057d1d-trusted-ca\") pod \"console-operator-58897d9998-62l7r\" (UID: \"32ca5dd6-f353-4c80-ad4b-1dacf3057d1d\") " pod="openshift-console-operator/console-operator-58897d9998-62l7r" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808097 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j8m72\" (UID: \"7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808122 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzwln\" (UniqueName: \"kubernetes.io/projected/40553430-c0eb-45b4-880b-d0e1f85adb33-kube-api-access-fzwln\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808169 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ba31bbad-4048-4651-92cd-d0cba1179541-node-pullsecrets\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808204 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/40553430-c0eb-45b4-880b-d0e1f85adb33-encryption-config\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808263 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5229f915-0d18-41e7-abaf-1d2a783670eb-auth-proxy-config\") pod \"machine-approver-56656f9798-75rsr\" (UID: \"5229f915-0d18-41e7-abaf-1d2a783670eb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808288 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ba31bbad-4048-4651-92cd-d0cba1179541-etcd-serving-ca\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808330 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5761a71-91cf-4781-9d7d-dc3f9b9096d0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-z4k8p\" (UID: \"b5761a71-91cf-4781-9d7d-dc3f9b9096d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808356 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/40553430-c0eb-45b4-880b-d0e1f85adb33-etcd-client\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808377 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/606e54dc-5cc4-4e02-b796-dd37b613a8d0-serving-cert\") pod \"controller-manager-879f6c89f-pddk8\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808424 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5229f915-0d18-41e7-abaf-1d2a783670eb-config\") pod \"machine-approver-56656f9798-75rsr\" (UID: \"5229f915-0d18-41e7-abaf-1d2a783670eb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808448 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-serving-cert\") pod \"route-controller-manager-6576b87f9c-f68ml\" (UID: \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808471 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ba31bbad-4048-4651-92cd-d0cba1179541-audit-dir\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808525 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrnzn\" (UniqueName: \"kubernetes.io/projected/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-kube-api-access-zrnzn\") pod \"route-controller-manager-6576b87f9c-f68ml\" (UID: \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808530 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808551 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbnvv\" (UniqueName: \"kubernetes.io/projected/32ca5dd6-f353-4c80-ad4b-1dacf3057d1d-kube-api-access-wbnvv\") pod \"console-operator-58897d9998-62l7r\" (UID: \"32ca5dd6-f353-4c80-ad4b-1dacf3057d1d\") " pod="openshift-console-operator/console-operator-58897d9998-62l7r" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808604 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/40553430-c0eb-45b4-880b-d0e1f85adb33-audit-policies\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808627 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-config\") pod \"controller-manager-879f6c89f-pddk8\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808681 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40553430-c0eb-45b4-880b-d0e1f85adb33-serving-cert\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808707 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32ca5dd6-f353-4c80-ad4b-1dacf3057d1d-serving-cert\") pod \"console-operator-58897d9998-62l7r\" (UID: \"32ca5dd6-f353-4c80-ad4b-1dacf3057d1d\") " pod="openshift-console-operator/console-operator-58897d9998-62l7r" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.808756 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ba31bbad-4048-4651-92cd-d0cba1179541-image-import-ca\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.809003 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ba31bbad-4048-4651-92cd-d0cba1179541-trusted-ca-bundle\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.809051 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/40553430-c0eb-45b4-880b-d0e1f85adb33-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.809090 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-config\") pod \"route-controller-manager-6576b87f9c-f68ml\" (UID: \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.810129 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40553430-c0eb-45b4-880b-d0e1f85adb33-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.810180 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53e674b8-9044-45f1-8aa5-bb81ec63ae4a-serving-cert\") pod \"openshift-config-operator-7777fb866f-jhptt\" (UID: \"53e674b8-9044-45f1-8aa5-bb81ec63ae4a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.810210 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6-images\") pod \"machine-api-operator-5694c8668f-j8m72\" (UID: \"7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.810236 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ba31bbad-4048-4651-92cd-d0cba1179541-audit\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.810252 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ba31bbad-4048-4651-92cd-d0cba1179541-encryption-config\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.810269 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzfk5\" (UniqueName: \"kubernetes.io/projected/ba31bbad-4048-4651-92cd-d0cba1179541-kube-api-access-nzfk5\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.810291 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-client-ca\") pod \"controller-manager-879f6c89f-pddk8\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.810425 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vrgcp"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.811360 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-z2mdj"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.811772 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.812112 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-vrgcp" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.812428 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.812809 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.813390 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-62l7r"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.814315 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.827857 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.829388 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.829675 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.829941 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.830215 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bl2l4"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.830465 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.830634 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.830727 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.830732 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.830778 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.831680 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.831716 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bl2l4" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.831727 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.833718 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.834113 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mzgjp"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.834690 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.835355 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.853076 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.854224 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.854556 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.860550 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.865514 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-tqxb8"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.866507 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-tqxb8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.868861 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.891792 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.892480 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.895290 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8nlkb"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.898986 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.902905 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916048 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c2d5j"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916081 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j8m72"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916095 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ln6xb"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916108 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8r5hx"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916119 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916134 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7m42k"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.913356 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-config\") pod \"route-controller-manager-6576b87f9c-f68ml\" (UID: \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.911876 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.912260 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-config\") pod \"route-controller-manager-6576b87f9c-f68ml\" (UID: \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916499 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40553430-c0eb-45b4-880b-d0e1f85adb33-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916531 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53e674b8-9044-45f1-8aa5-bb81ec63ae4a-serving-cert\") pod \"openshift-config-operator-7777fb866f-jhptt\" (UID: \"53e674b8-9044-45f1-8aa5-bb81ec63ae4a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916562 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6-images\") pod \"machine-api-operator-5694c8668f-j8m72\" (UID: \"7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916587 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzfk5\" (UniqueName: \"kubernetes.io/projected/ba31bbad-4048-4651-92cd-d0cba1179541-kube-api-access-nzfk5\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916618 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ba31bbad-4048-4651-92cd-d0cba1179541-audit\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916641 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ba31bbad-4048-4651-92cd-d0cba1179541-encryption-config\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916668 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-client-ca\") pod \"controller-manager-879f6c89f-pddk8\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916702 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b4637a3-8d20-4427-8e71-9b5ab8e6ee8c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8nlkb\" (UID: \"5b4637a3-8d20-4427-8e71-9b5ab8e6ee8c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8nlkb" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916728 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5761a71-91cf-4781-9d7d-dc3f9b9096d0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-z4k8p\" (UID: \"b5761a71-91cf-4781-9d7d-dc3f9b9096d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916767 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6-config\") pod \"machine-api-operator-5694c8668f-j8m72\" (UID: \"7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916792 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w9cv\" (UniqueName: \"kubernetes.io/projected/7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6-kube-api-access-7w9cv\") pod \"machine-api-operator-5694c8668f-j8m72\" (UID: \"7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916812 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5229f915-0d18-41e7-abaf-1d2a783670eb-machine-approver-tls\") pod \"machine-approver-56656f9798-75rsr\" (UID: \"5229f915-0d18-41e7-abaf-1d2a783670eb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916836 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-946xt\" (UniqueName: \"kubernetes.io/projected/b5761a71-91cf-4781-9d7d-dc3f9b9096d0-kube-api-access-946xt\") pod \"openshift-apiserver-operator-796bbdcf4f-z4k8p\" (UID: \"b5761a71-91cf-4781-9d7d-dc3f9b9096d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916864 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/53e674b8-9044-45f1-8aa5-bb81ec63ae4a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-jhptt\" (UID: \"53e674b8-9044-45f1-8aa5-bb81ec63ae4a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916901 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj52k\" (UniqueName: \"kubernetes.io/projected/5b4637a3-8d20-4427-8e71-9b5ab8e6ee8c-kube-api-access-gj52k\") pod \"cluster-samples-operator-665b6dd947-8nlkb\" (UID: \"5b4637a3-8d20-4427-8e71-9b5ab8e6ee8c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8nlkb" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916950 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc5dm\" (UniqueName: \"kubernetes.io/projected/53e674b8-9044-45f1-8aa5-bb81ec63ae4a-kube-api-access-nc5dm\") pod \"openshift-config-operator-7777fb866f-jhptt\" (UID: \"53e674b8-9044-45f1-8aa5-bb81ec63ae4a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916975 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s46kq\" (UniqueName: \"kubernetes.io/projected/606e54dc-5cc4-4e02-b796-dd37b613a8d0-kube-api-access-s46kq\") pod \"controller-manager-879f6c89f-pddk8\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.916995 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba31bbad-4048-4651-92cd-d0cba1179541-serving-cert\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917018 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba31bbad-4048-4651-92cd-d0cba1179541-config\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917038 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ba31bbad-4048-4651-92cd-d0cba1179541-etcd-client\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917059 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-client-ca\") pod \"route-controller-manager-6576b87f9c-f68ml\" (UID: \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917078 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ca5dd6-f353-4c80-ad4b-1dacf3057d1d-config\") pod \"console-operator-58897d9998-62l7r\" (UID: \"32ca5dd6-f353-4c80-ad4b-1dacf3057d1d\") " pod="openshift-console-operator/console-operator-58897d9998-62l7r" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917094 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-pddk8\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917109 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/40553430-c0eb-45b4-880b-d0e1f85adb33-audit-dir\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917137 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zj6n\" (UniqueName: \"kubernetes.io/projected/5229f915-0d18-41e7-abaf-1d2a783670eb-kube-api-access-9zj6n\") pod \"machine-approver-56656f9798-75rsr\" (UID: \"5229f915-0d18-41e7-abaf-1d2a783670eb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917153 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j8m72\" (UID: \"7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917171 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/32ca5dd6-f353-4c80-ad4b-1dacf3057d1d-trusted-ca\") pod \"console-operator-58897d9998-62l7r\" (UID: \"32ca5dd6-f353-4c80-ad4b-1dacf3057d1d\") " pod="openshift-console-operator/console-operator-58897d9998-62l7r" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917190 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzwln\" (UniqueName: \"kubernetes.io/projected/40553430-c0eb-45b4-880b-d0e1f85adb33-kube-api-access-fzwln\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917207 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ba31bbad-4048-4651-92cd-d0cba1179541-node-pullsecrets\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917227 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/40553430-c0eb-45b4-880b-d0e1f85adb33-encryption-config\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917243 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5761a71-91cf-4781-9d7d-dc3f9b9096d0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-z4k8p\" (UID: \"b5761a71-91cf-4781-9d7d-dc3f9b9096d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917261 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5229f915-0d18-41e7-abaf-1d2a783670eb-auth-proxy-config\") pod \"machine-approver-56656f9798-75rsr\" (UID: \"5229f915-0d18-41e7-abaf-1d2a783670eb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917277 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ba31bbad-4048-4651-92cd-d0cba1179541-etcd-serving-ca\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917296 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/40553430-c0eb-45b4-880b-d0e1f85adb33-etcd-client\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917313 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/606e54dc-5cc4-4e02-b796-dd37b613a8d0-serving-cert\") pod \"controller-manager-879f6c89f-pddk8\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917349 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5229f915-0d18-41e7-abaf-1d2a783670eb-config\") pod \"machine-approver-56656f9798-75rsr\" (UID: \"5229f915-0d18-41e7-abaf-1d2a783670eb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917366 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-serving-cert\") pod \"route-controller-manager-6576b87f9c-f68ml\" (UID: \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917381 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ba31bbad-4048-4651-92cd-d0cba1179541-audit-dir\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917398 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrnzn\" (UniqueName: \"kubernetes.io/projected/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-kube-api-access-zrnzn\") pod \"route-controller-manager-6576b87f9c-f68ml\" (UID: \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917422 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbnvv\" (UniqueName: \"kubernetes.io/projected/32ca5dd6-f353-4c80-ad4b-1dacf3057d1d-kube-api-access-wbnvv\") pod \"console-operator-58897d9998-62l7r\" (UID: \"32ca5dd6-f353-4c80-ad4b-1dacf3057d1d\") " pod="openshift-console-operator/console-operator-58897d9998-62l7r" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917447 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/40553430-c0eb-45b4-880b-d0e1f85adb33-audit-policies\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917464 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-config\") pod \"controller-manager-879f6c89f-pddk8\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917484 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40553430-c0eb-45b4-880b-d0e1f85adb33-serving-cert\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917498 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32ca5dd6-f353-4c80-ad4b-1dacf3057d1d-serving-cert\") pod \"console-operator-58897d9998-62l7r\" (UID: \"32ca5dd6-f353-4c80-ad4b-1dacf3057d1d\") " pod="openshift-console-operator/console-operator-58897d9998-62l7r" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917513 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ba31bbad-4048-4651-92cd-d0cba1179541-image-import-ca\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917540 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ba31bbad-4048-4651-92cd-d0cba1179541-trusted-ca-bundle\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.917560 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/40553430-c0eb-45b4-880b-d0e1f85adb33-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.920624 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/40553430-c0eb-45b4-880b-d0e1f85adb33-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.921313 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-client-ca\") pod \"route-controller-manager-6576b87f9c-f68ml\" (UID: \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.921498 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40553430-c0eb-45b4-880b-d0e1f85adb33-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.924072 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ba31bbad-4048-4651-92cd-d0cba1179541-audit\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.927788 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6-images\") pod \"machine-api-operator-5694c8668f-j8m72\" (UID: \"7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.928108 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/53e674b8-9044-45f1-8aa5-bb81ec63ae4a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-jhptt\" (UID: \"53e674b8-9044-45f1-8aa5-bb81ec63ae4a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.928435 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/40553430-c0eb-45b4-880b-d0e1f85adb33-audit-policies\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.928882 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5229f915-0d18-41e7-abaf-1d2a783670eb-config\") pod \"machine-approver-56656f9798-75rsr\" (UID: \"5229f915-0d18-41e7-abaf-1d2a783670eb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.929485 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ba31bbad-4048-4651-92cd-d0cba1179541-audit-dir\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.929495 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ca5dd6-f353-4c80-ad4b-1dacf3057d1d-config\") pod \"console-operator-58897d9998-62l7r\" (UID: \"32ca5dd6-f353-4c80-ad4b-1dacf3057d1d\") " pod="openshift-console-operator/console-operator-58897d9998-62l7r" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.930472 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53e674b8-9044-45f1-8aa5-bb81ec63ae4a-serving-cert\") pod \"openshift-config-operator-7777fb866f-jhptt\" (UID: \"53e674b8-9044-45f1-8aa5-bb81ec63ae4a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.931117 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-pddk8\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.931286 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/40553430-c0eb-45b4-880b-d0e1f85adb33-audit-dir\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.933575 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5229f915-0d18-41e7-abaf-1d2a783670eb-machine-approver-tls\") pod \"machine-approver-56656f9798-75rsr\" (UID: \"5229f915-0d18-41e7-abaf-1d2a783670eb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.934140 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.936529 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba31bbad-4048-4651-92cd-d0cba1179541-config\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.937258 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/32ca5dd6-f353-4c80-ad4b-1dacf3057d1d-trusted-ca\") pod \"console-operator-58897d9998-62l7r\" (UID: \"32ca5dd6-f353-4c80-ad4b-1dacf3057d1d\") " pod="openshift-console-operator/console-operator-58897d9998-62l7r" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.937740 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vrgcp"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.937792 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.938793 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ba31bbad-4048-4651-92cd-d0cba1179541-image-import-ca\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.938842 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.939869 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5761a71-91cf-4781-9d7d-dc3f9b9096d0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-z4k8p\" (UID: \"b5761a71-91cf-4781-9d7d-dc3f9b9096d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.940332 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5229f915-0d18-41e7-abaf-1d2a783670eb-auth-proxy-config\") pod \"machine-approver-56656f9798-75rsr\" (UID: \"5229f915-0d18-41e7-abaf-1d2a783670eb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.940709 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ba31bbad-4048-4651-92cd-d0cba1179541-etcd-serving-ca\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.941187 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ba31bbad-4048-4651-92cd-d0cba1179541-node-pullsecrets\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.943146 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-config\") pod \"controller-manager-879f6c89f-pddk8\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.944533 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.944571 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-7lbq7"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.944583 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6qzsg"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.944614 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-client-ca\") pod \"controller-manager-879f6c89f-pddk8\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.944893 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ba31bbad-4048-4651-92cd-d0cba1179541-trusted-ca-bundle\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.945270 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6-config\") pod \"machine-api-operator-5694c8668f-j8m72\" (UID: \"7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.945432 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba31bbad-4048-4651-92cd-d0cba1179541-serving-cert\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.945603 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j8m72\" (UID: \"7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.946317 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.947337 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ba31bbad-4048-4651-92cd-d0cba1179541-etcd-client\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.947631 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-tclrk"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.947867 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40553430-c0eb-45b4-880b-d0e1f85adb33-serving-cert\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.948225 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.948263 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-tclrk" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.949068 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cj594"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.950392 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-zp567"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.950631 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.951360 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-zp567" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.954955 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/40553430-c0eb-45b4-880b-d0e1f85adb33-encryption-config\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.955150 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5761a71-91cf-4781-9d7d-dc3f9b9096d0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-z4k8p\" (UID: \"b5761a71-91cf-4781-9d7d-dc3f9b9096d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.955333 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ba31bbad-4048-4651-92cd-d0cba1179541-encryption-config\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.952046 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-z2mdj"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.955498 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/40553430-c0eb-45b4-880b-d0e1f85adb33-etcd-client\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.955808 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-serving-cert\") pod \"route-controller-manager-6576b87f9c-f68ml\" (UID: \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.956085 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32ca5dd6-f353-4c80-ad4b-1dacf3057d1d-serving-cert\") pod \"console-operator-58897d9998-62l7r\" (UID: \"32ca5dd6-f353-4c80-ad4b-1dacf3057d1d\") " pod="openshift-console-operator/console-operator-58897d9998-62l7r" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.959037 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.959085 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.960534 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.962485 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/606e54dc-5cc4-4e02-b796-dd37b613a8d0-serving-cert\") pod \"controller-manager-879f6c89f-pddk8\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.962560 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.964158 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.965774 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mzgjp"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.968217 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.969159 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dgkrj"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.970344 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.971611 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.972963 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.974100 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fd998"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.975376 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pddk8"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.976588 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cj594"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.977532 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.979182 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bl2l4"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.979581 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.980401 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-tqxb8"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.981652 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5b4637a3-8d20-4427-8e71-9b5ab8e6ee8c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8nlkb\" (UID: \"5b4637a3-8d20-4427-8e71-9b5ab8e6ee8c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8nlkb" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.981705 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-zp567"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.983751 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-f47vx"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.984658 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-f47vx"] Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.984773 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-f47vx" Dec 04 03:56:59 crc kubenswrapper[4806]: I1204 03:56:59.987843 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.007750 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.027896 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.047866 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.068231 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.088058 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.109303 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.128177 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.148707 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.167729 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.208642 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.228350 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.247832 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.268495 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.288498 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.308968 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.328183 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.348967 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.369087 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.388032 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.408224 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.428559 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.447767 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.468974 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.488849 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.509262 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.528257 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.555402 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.568815 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.589537 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.628416 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.648157 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.668976 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.688760 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.709435 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.728579 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.747860 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.774774 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.787938 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.809393 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.826734 4806 request.go:700] Waited for 1.013664887s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager-operator/secrets?fieldSelector=metadata.name%3Dkube-controller-manager-operator-dockercfg-gkqpw&limit=500&resourceVersion=0 Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.829053 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.849179 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.868401 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.889284 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.908507 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.929784 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.948978 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.968572 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 04 03:57:00 crc kubenswrapper[4806]: I1204 03:57:00.988086 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.009522 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.030063 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.049184 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.068008 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.089008 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.108692 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.128855 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.149410 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.168467 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.197376 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.207648 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.229530 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.248780 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.268958 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.289300 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.308667 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.327756 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.348499 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.368443 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.389224 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.409455 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.428803 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.448340 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.483037 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj52k\" (UniqueName: \"kubernetes.io/projected/5b4637a3-8d20-4427-8e71-9b5ab8e6ee8c-kube-api-access-gj52k\") pod \"cluster-samples-operator-665b6dd947-8nlkb\" (UID: \"5b4637a3-8d20-4427-8e71-9b5ab8e6ee8c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8nlkb" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.503660 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w9cv\" (UniqueName: \"kubernetes.io/projected/7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6-kube-api-access-7w9cv\") pod \"machine-api-operator-5694c8668f-j8m72\" (UID: \"7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.526031 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzfk5\" (UniqueName: \"kubernetes.io/projected/ba31bbad-4048-4651-92cd-d0cba1179541-kube-api-access-nzfk5\") pod \"apiserver-76f77b778f-7lbq7\" (UID: \"ba31bbad-4048-4651-92cd-d0cba1179541\") " pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.543058 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc5dm\" (UniqueName: \"kubernetes.io/projected/53e674b8-9044-45f1-8aa5-bb81ec63ae4a-kube-api-access-nc5dm\") pod \"openshift-config-operator-7777fb866f-jhptt\" (UID: \"53e674b8-9044-45f1-8aa5-bb81ec63ae4a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.564419 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.568063 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s46kq\" (UniqueName: \"kubernetes.io/projected/606e54dc-5cc4-4e02-b796-dd37b613a8d0-kube-api-access-s46kq\") pod \"controller-manager-879f6c89f-pddk8\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.582606 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-946xt\" (UniqueName: \"kubernetes.io/projected/b5761a71-91cf-4781-9d7d-dc3f9b9096d0-kube-api-access-946xt\") pod \"openshift-apiserver-operator-796bbdcf4f-z4k8p\" (UID: \"b5761a71-91cf-4781-9d7d-dc3f9b9096d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.602357 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.603906 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrnzn\" (UniqueName: \"kubernetes.io/projected/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-kube-api-access-zrnzn\") pod \"route-controller-manager-6576b87f9c-f68ml\" (UID: \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.616856 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8nlkb" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.625788 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbnvv\" (UniqueName: \"kubernetes.io/projected/32ca5dd6-f353-4c80-ad4b-1dacf3057d1d-kube-api-access-wbnvv\") pod \"console-operator-58897d9998-62l7r\" (UID: \"32ca5dd6-f353-4c80-ad4b-1dacf3057d1d\") " pod="openshift-console-operator/console-operator-58897d9998-62l7r" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.644566 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zj6n\" (UniqueName: \"kubernetes.io/projected/5229f915-0d18-41e7-abaf-1d2a783670eb-kube-api-access-9zj6n\") pod \"machine-approver-56656f9798-75rsr\" (UID: \"5229f915-0d18-41e7-abaf-1d2a783670eb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.657381 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.664507 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzwln\" (UniqueName: \"kubernetes.io/projected/40553430-c0eb-45b4-880b-d0e1f85adb33-kube-api-access-fzwln\") pod \"apiserver-7bbb656c7d-g89mj\" (UID: \"40553430-c0eb-45b4-880b-d0e1f85adb33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.670710 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.713236 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.713520 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.716906 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.717034 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.728666 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.748042 4806 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.768847 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.789185 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.809022 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.809164 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.826526 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.827783 4806 request.go:700] Waited for 1.876186686s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Dcanary-serving-cert&limit=500&resourceVersion=0 Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.830546 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.837393 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-62l7r" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.850203 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.868130 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.888395 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.909727 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.921265 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" Dec 04 03:57:01 crc kubenswrapper[4806]: I1204 03:57:01.957269 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8nlkb"] Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.042494 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.042526 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f43cb7f-513b-4958-88f6-e43a01c7adda-trusted-ca\") pod \"ingress-operator-5b745b69d9-mgzl9\" (UID: \"9f43cb7f-513b-4958-88f6-e43a01c7adda\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.042545 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f43cb7f-513b-4958-88f6-e43a01c7adda-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mgzl9\" (UID: \"9f43cb7f-513b-4958-88f6-e43a01c7adda\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.042561 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/000c6858-a809-46e4-bdcb-b150477e5ce2-audit-dir\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.043878 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.043938 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-registry-tls\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.043978 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/920cdae1-9503-4181-92a2-d62969908d1d-proxy-tls\") pod \"machine-config-operator-74547568cd-9dhq6\" (UID: \"920cdae1-9503-4181-92a2-d62969908d1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044011 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/861e4822-a809-46c7-af83-06e8978ca8aa-stats-auth\") pod \"router-default-5444994796-mmxrh\" (UID: \"861e4822-a809-46c7-af83-06e8978ca8aa\") " pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044038 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044072 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9f43cb7f-513b-4958-88f6-e43a01c7adda-metrics-tls\") pod \"ingress-operator-5b745b69d9-mgzl9\" (UID: \"9f43cb7f-513b-4958-88f6-e43a01c7adda\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044095 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/38a0c15f-4422-4119-b513-2e0ede6e8e8a-srv-cert\") pod \"olm-operator-6b444d44fb-2c4tv\" (UID: \"38a0c15f-4422-4119-b513-2e0ede6e8e8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044140 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/79fbf847-050b-404d-8e9a-7f16f42089dc-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-62ppz\" (UID: \"79fbf847-050b-404d-8e9a-7f16f42089dc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044160 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd2nl\" (UniqueName: \"kubernetes.io/projected/79fbf847-050b-404d-8e9a-7f16f42089dc-kube-api-access-qd2nl\") pod \"cluster-image-registry-operator-dc59b4c8b-62ppz\" (UID: \"79fbf847-050b-404d-8e9a-7f16f42089dc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044196 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/920cdae1-9503-4181-92a2-d62969908d1d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9dhq6\" (UID: \"920cdae1-9503-4181-92a2-d62969908d1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044216 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75hqw\" (UniqueName: \"kubernetes.io/projected/9f43cb7f-513b-4958-88f6-e43a01c7adda-kube-api-access-75hqw\") pod \"ingress-operator-5b745b69d9-mgzl9\" (UID: \"9f43cb7f-513b-4958-88f6-e43a01c7adda\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044234 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/79fbf847-050b-404d-8e9a-7f16f42089dc-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-62ppz\" (UID: \"79fbf847-050b-404d-8e9a-7f16f42089dc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044490 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/861e4822-a809-46c7-af83-06e8978ca8aa-metrics-certs\") pod \"router-default-5444994796-mmxrh\" (UID: \"861e4822-a809-46c7-af83-06e8978ca8aa\") " pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044532 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqxmk\" (UniqueName: \"kubernetes.io/projected/e9ff0349-32de-4d48-87ad-3e206ec56b2e-kube-api-access-nqxmk\") pod \"authentication-operator-69f744f599-6qzsg\" (UID: \"e9ff0349-32de-4d48-87ad-3e206ec56b2e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044569 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9ff0349-32de-4d48-87ad-3e206ec56b2e-config\") pod \"authentication-operator-69f744f599-6qzsg\" (UID: \"e9ff0349-32de-4d48-87ad-3e206ec56b2e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044594 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/233ad320-ca2f-40bf-bdcf-64e5aed5509f-trusted-ca\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044614 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044692 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hpkz\" (UniqueName: \"kubernetes.io/projected/920cdae1-9503-4181-92a2-d62969908d1d-kube-api-access-2hpkz\") pod \"machine-config-operator-74547568cd-9dhq6\" (UID: \"920cdae1-9503-4181-92a2-d62969908d1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044729 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/0647d456-31a3-4906-9d62-f43b4cfef5cb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dgkrj\" (UID: \"0647d456-31a3-4906-9d62-f43b4cfef5cb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dgkrj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044806 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c85a9a8a-89cf-4c60-9619-879f64d40195-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsqzh\" (UID: \"c85a9a8a-89cf-4c60-9619-879f64d40195\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044865 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/233ad320-ca2f-40bf-bdcf-64e5aed5509f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044888 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/233ad320-ca2f-40bf-bdcf-64e5aed5509f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044907 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70d37d83-4fab-49db-8cb5-b98f50c02a0e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-db5db\" (UID: \"70d37d83-4fab-49db-8cb5-b98f50c02a0e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044944 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.044977 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9ff0349-32de-4d48-87ad-3e206ec56b2e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6qzsg\" (UID: \"e9ff0349-32de-4d48-87ad-3e206ec56b2e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.045039 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.045520 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70d37d83-4fab-49db-8cb5-b98f50c02a0e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-db5db\" (UID: \"70d37d83-4fab-49db-8cb5-b98f50c02a0e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.045543 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e-config\") pod \"kube-apiserver-operator-766d6c64bb-djtqn\" (UID: \"7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.045564 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5jq4\" (UniqueName: \"kubernetes.io/projected/46000442-50bc-4480-9939-6a56032d3d2d-kube-api-access-w5jq4\") pod \"downloads-7954f5f757-8r5hx\" (UID: \"46000442-50bc-4480-9939-6a56032d3d2d\") " pod="openshift-console/downloads-7954f5f757-8r5hx" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.045636 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.045670 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/861e4822-a809-46c7-af83-06e8978ca8aa-service-ca-bundle\") pod \"router-default-5444994796-mmxrh\" (UID: \"861e4822-a809-46c7-af83-06e8978ca8aa\") " pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.045690 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9ff0349-32de-4d48-87ad-3e206ec56b2e-service-ca-bundle\") pod \"authentication-operator-69f744f599-6qzsg\" (UID: \"e9ff0349-32de-4d48-87ad-3e206ec56b2e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.045709 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d10811c5-1f44-4668-a5f8-9690f2d81983-etcd-client\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.045742 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/861e4822-a809-46c7-af83-06e8978ca8aa-default-certificate\") pod \"router-default-5444994796-mmxrh\" (UID: \"861e4822-a809-46c7-af83-06e8978ca8aa\") " pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.045761 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmfkm\" (UniqueName: \"kubernetes.io/projected/0647d456-31a3-4906-9d62-f43b4cfef5cb-kube-api-access-mmfkm\") pod \"control-plane-machine-set-operator-78cbb6b69f-dgkrj\" (UID: \"0647d456-31a3-4906-9d62-f43b4cfef5cb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dgkrj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.045825 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d10811c5-1f44-4668-a5f8-9690f2d81983-serving-cert\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.045846 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxnc5\" (UniqueName: \"kubernetes.io/projected/70d37d83-4fab-49db-8cb5-b98f50c02a0e-kube-api-access-vxnc5\") pod \"openshift-controller-manager-operator-756b6f6bc6-db5db\" (UID: \"70d37d83-4fab-49db-8cb5-b98f50c02a0e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.045865 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-djtqn\" (UID: \"7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.045964 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmlrq\" (UniqueName: \"kubernetes.io/projected/d10811c5-1f44-4668-a5f8-9690f2d81983-kube-api-access-gmlrq\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.045991 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3b036b1c-d79d-49d5-89ab-c669a5f8f47b-metrics-tls\") pod \"dns-operator-744455d44c-ln6xb\" (UID: \"3b036b1c-d79d-49d5-89ab-c669a5f8f47b\") " pod="openshift-dns-operator/dns-operator-744455d44c-ln6xb" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046152 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c85a9a8a-89cf-4c60-9619-879f64d40195-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsqzh\" (UID: \"c85a9a8a-89cf-4c60-9619-879f64d40195\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046197 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046217 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046250 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/79fbf847-050b-404d-8e9a-7f16f42089dc-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-62ppz\" (UID: \"79fbf847-050b-404d-8e9a-7f16f42089dc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046301 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkb2v\" (UniqueName: \"kubernetes.io/projected/c85a9a8a-89cf-4c60-9619-879f64d40195-kube-api-access-rkb2v\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsqzh\" (UID: \"c85a9a8a-89cf-4c60-9619-879f64d40195\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046372 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6c5j\" (UniqueName: \"kubernetes.io/projected/000c6858-a809-46e4-bdcb-b150477e5ce2-kube-api-access-f6c5j\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046417 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmv2r\" (UniqueName: \"kubernetes.io/projected/861e4822-a809-46c7-af83-06e8978ca8aa-kube-api-access-qmv2r\") pod \"router-default-5444994796-mmxrh\" (UID: \"861e4822-a809-46c7-af83-06e8978ca8aa\") " pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046437 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mmqr\" (UniqueName: \"kubernetes.io/projected/3b036b1c-d79d-49d5-89ab-c669a5f8f47b-kube-api-access-7mmqr\") pod \"dns-operator-744455d44c-ln6xb\" (UID: \"3b036b1c-d79d-49d5-89ab-c669a5f8f47b\") " pod="openshift-dns-operator/dns-operator-744455d44c-ln6xb" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046456 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d10811c5-1f44-4668-a5f8-9690f2d81983-config\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046569 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d10811c5-1f44-4668-a5f8-9690f2d81983-etcd-ca\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046624 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046657 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9ff0349-32de-4d48-87ad-3e206ec56b2e-serving-cert\") pod \"authentication-operator-69f744f599-6qzsg\" (UID: \"e9ff0349-32de-4d48-87ad-3e206ec56b2e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046677 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046821 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-djtqn\" (UID: \"7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046853 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dfts\" (UniqueName: \"kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-kube-api-access-4dfts\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046885 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.046904 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/920cdae1-9503-4181-92a2-d62969908d1d-images\") pod \"machine-config-operator-74547568cd-9dhq6\" (UID: \"920cdae1-9503-4181-92a2-d62969908d1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.047150 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/233ad320-ca2f-40bf-bdcf-64e5aed5509f-registry-certificates\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.047181 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-bound-sa-token\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.047335 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-audit-policies\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.047374 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d10811c5-1f44-4668-a5f8-9690f2d81983-etcd-service-ca\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.047407 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/38a0c15f-4422-4119-b513-2e0ede6e8e8a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2c4tv\" (UID: \"38a0c15f-4422-4119-b513-2e0ede6e8e8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.047442 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnp87\" (UniqueName: \"kubernetes.io/projected/38a0c15f-4422-4119-b513-2e0ede6e8e8a-kube-api-access-bnp87\") pod \"olm-operator-6b444d44fb-2c4tv\" (UID: \"38a0c15f-4422-4119-b513-2e0ede6e8e8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv" Dec 04 03:57:02 crc kubenswrapper[4806]: E1204 03:57:02.051449 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:02.551433484 +0000 UTC m=+137.409946412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.057827 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" event={"ID":"5229f915-0d18-41e7-abaf-1d2a783670eb","Type":"ContainerStarted","Data":"9a32862ed67652c12018202a0f464108615bd1f651c3ebe0a4dd7e9040807daf"} Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.093777 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j8m72"] Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.099402 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-7lbq7"] Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.148500 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.148740 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/38a0c15f-4422-4119-b513-2e0ede6e8e8a-srv-cert\") pod \"olm-operator-6b444d44fb-2c4tv\" (UID: \"38a0c15f-4422-4119-b513-2e0ede6e8e8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.148773 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/79fbf847-050b-404d-8e9a-7f16f42089dc-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-62ppz\" (UID: \"79fbf847-050b-404d-8e9a-7f16f42089dc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.148796 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd2nl\" (UniqueName: \"kubernetes.io/projected/79fbf847-050b-404d-8e9a-7f16f42089dc-kube-api-access-qd2nl\") pod \"cluster-image-registry-operator-dc59b4c8b-62ppz\" (UID: \"79fbf847-050b-404d-8e9a-7f16f42089dc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.148821 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-csi-data-dir\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.148842 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1905e8ad-dfcf-4ce5-9565-28562f0832bc-config\") pod \"kube-controller-manager-operator-78b949d7b-2fq8f\" (UID: \"1905e8ad-dfcf-4ce5-9565-28562f0832bc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.148898 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/920cdae1-9503-4181-92a2-d62969908d1d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9dhq6\" (UID: \"920cdae1-9503-4181-92a2-d62969908d1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.148937 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75hqw\" (UniqueName: \"kubernetes.io/projected/9f43cb7f-513b-4958-88f6-e43a01c7adda-kube-api-access-75hqw\") pod \"ingress-operator-5b745b69d9-mgzl9\" (UID: \"9f43cb7f-513b-4958-88f6-e43a01c7adda\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.148958 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/79fbf847-050b-404d-8e9a-7f16f42089dc-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-62ppz\" (UID: \"79fbf847-050b-404d-8e9a-7f16f42089dc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.148980 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv6mt\" (UniqueName: \"kubernetes.io/projected/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-kube-api-access-gv6mt\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149016 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9ff0349-32de-4d48-87ad-3e206ec56b2e-config\") pod \"authentication-operator-69f744f599-6qzsg\" (UID: \"e9ff0349-32de-4d48-87ad-3e206ec56b2e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149040 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e98c9985-31a9-4f14-818d-4519f0a3131b-signing-key\") pod \"service-ca-9c57cc56f-tqxb8\" (UID: \"e98c9985-31a9-4f14-818d-4519f0a3131b\") " pod="openshift-service-ca/service-ca-9c57cc56f-tqxb8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149065 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/233ad320-ca2f-40bf-bdcf-64e5aed5509f-trusted-ca\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149087 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hpkz\" (UniqueName: \"kubernetes.io/projected/920cdae1-9503-4181-92a2-d62969908d1d-kube-api-access-2hpkz\") pod \"machine-config-operator-74547568cd-9dhq6\" (UID: \"920cdae1-9503-4181-92a2-d62969908d1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149112 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/0647d456-31a3-4906-9d62-f43b4cfef5cb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dgkrj\" (UID: \"0647d456-31a3-4906-9d62-f43b4cfef5cb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dgkrj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149141 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-plugins-dir\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149162 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-socket-dir\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149190 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/233ad320-ca2f-40bf-bdcf-64e5aed5509f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149222 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70d37d83-4fab-49db-8cb5-b98f50c02a0e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-db5db\" (UID: \"70d37d83-4fab-49db-8cb5-b98f50c02a0e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149247 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-service-ca\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149286 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149311 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5fa998f-665b-421d-b589-e2531df3b2c2-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cvm28\" (UID: \"f5fa998f-665b-421d-b589-e2531df3b2c2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149336 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/957c4786-76f4-4437-b56a-d42dee2c65b5-certs\") pod \"machine-config-server-tclrk\" (UID: \"957c4786-76f4-4437-b56a-d42dee2c65b5\") " pod="openshift-machine-config-operator/machine-config-server-tclrk" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149359 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/08021692-e26d-42ab-85a2-136b1bd2b4ed-proxy-tls\") pod \"machine-config-controller-84d6567774-wbb9n\" (UID: \"08021692-e26d-42ab-85a2-136b1bd2b4ed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149383 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5jq4\" (UniqueName: \"kubernetes.io/projected/46000442-50bc-4480-9939-6a56032d3d2d-kube-api-access-w5jq4\") pod \"downloads-7954f5f757-8r5hx\" (UID: \"46000442-50bc-4480-9939-6a56032d3d2d\") " pod="openshift-console/downloads-7954f5f757-8r5hx" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149405 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-registration-dir\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: E1204 03:57:02.149761 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:02.649743161 +0000 UTC m=+137.508256089 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.154758 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/920cdae1-9503-4181-92a2-d62969908d1d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9dhq6\" (UID: \"920cdae1-9503-4181-92a2-d62969908d1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.155614 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/79fbf847-050b-404d-8e9a-7f16f42089dc-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-62ppz\" (UID: \"79fbf847-050b-404d-8e9a-7f16f42089dc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.155642 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9ff0349-32de-4d48-87ad-3e206ec56b2e-config\") pod \"authentication-operator-69f744f599-6qzsg\" (UID: \"e9ff0349-32de-4d48-87ad-3e206ec56b2e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.149450 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/861e4822-a809-46c7-af83-06e8978ca8aa-service-ca-bundle\") pod \"router-default-5444994796-mmxrh\" (UID: \"861e4822-a809-46c7-af83-06e8978ca8aa\") " pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.156530 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9ff0349-32de-4d48-87ad-3e206ec56b2e-service-ca-bundle\") pod \"authentication-operator-69f744f599-6qzsg\" (UID: \"e9ff0349-32de-4d48-87ad-3e206ec56b2e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.156580 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmfkm\" (UniqueName: \"kubernetes.io/projected/0647d456-31a3-4906-9d62-f43b4cfef5cb-kube-api-access-mmfkm\") pod \"control-plane-machine-set-operator-78cbb6b69f-dgkrj\" (UID: \"0647d456-31a3-4906-9d62-f43b4cfef5cb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dgkrj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.156885 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxnc5\" (UniqueName: \"kubernetes.io/projected/70d37d83-4fab-49db-8cb5-b98f50c02a0e-kube-api-access-vxnc5\") pod \"openshift-controller-manager-operator-756b6f6bc6-db5db\" (UID: \"70d37d83-4fab-49db-8cb5-b98f50c02a0e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.156909 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-djtqn\" (UID: \"7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.156956 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmlrq\" (UniqueName: \"kubernetes.io/projected/d10811c5-1f44-4668-a5f8-9690f2d81983-kube-api-access-gmlrq\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.156978 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cdcade7f-c50d-4690-a339-db8dcc9a7ef0-metrics-tls\") pod \"dns-default-f47vx\" (UID: \"cdcade7f-c50d-4690-a339-db8dcc9a7ef0\") " pod="openshift-dns/dns-default-f47vx" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.156994 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5m5h\" (UniqueName: \"kubernetes.io/projected/4e4767a8-ff51-43c7-94cc-625c9b8689a3-kube-api-access-j5m5h\") pod \"marketplace-operator-79b997595-mzgjp\" (UID: \"4e4767a8-ff51-43c7-94cc-625c9b8689a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157035 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157056 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157074 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/79fbf847-050b-404d-8e9a-7f16f42089dc-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-62ppz\" (UID: \"79fbf847-050b-404d-8e9a-7f16f42089dc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157113 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc5f8\" (UniqueName: \"kubernetes.io/projected/957c4786-76f4-4437-b56a-d42dee2c65b5-kube-api-access-tc5f8\") pod \"machine-config-server-tclrk\" (UID: \"957c4786-76f4-4437-b56a-d42dee2c65b5\") " pod="openshift-machine-config-operator/machine-config-server-tclrk" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157142 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcvgf\" (UniqueName: \"kubernetes.io/projected/a3f377d2-10db-41db-9361-06abff2e0d29-kube-api-access-lcvgf\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157173 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6c5j\" (UniqueName: \"kubernetes.io/projected/000c6858-a809-46e4-bdcb-b150477e5ce2-kube-api-access-f6c5j\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157193 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157211 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157251 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzjbm\" (UniqueName: \"kubernetes.io/projected/2f20b247-6932-459d-a19e-fd200ba45136-kube-api-access-zzjbm\") pod \"migrator-59844c95c7-bl2l4\" (UID: \"2f20b247-6932-459d-a19e-fd200ba45136\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bl2l4" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157274 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gll4j\" (UniqueName: \"kubernetes.io/projected/cdcade7f-c50d-4690-a339-db8dcc9a7ef0-kube-api-access-gll4j\") pod \"dns-default-f47vx\" (UID: \"cdcade7f-c50d-4690-a339-db8dcc9a7ef0\") " pod="openshift-dns/dns-default-f47vx" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157290 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1905e8ad-dfcf-4ce5-9565-28562f0832bc-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2fq8f\" (UID: \"1905e8ad-dfcf-4ce5-9565-28562f0832bc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157310 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157330 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/920cdae1-9503-4181-92a2-d62969908d1d-images\") pod \"machine-config-operator-74547568cd-9dhq6\" (UID: \"920cdae1-9503-4181-92a2-d62969908d1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157350 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-bound-sa-token\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157372 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-audit-policies\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157389 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d10811c5-1f44-4668-a5f8-9690f2d81983-etcd-service-ca\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157405 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e98c9985-31a9-4f14-818d-4519f0a3131b-signing-cabundle\") pod \"service-ca-9c57cc56f-tqxb8\" (UID: \"e98c9985-31a9-4f14-818d-4519f0a3131b\") " pod="openshift-service-ca/service-ca-9c57cc56f-tqxb8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157424 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnp87\" (UniqueName: \"kubernetes.io/projected/38a0c15f-4422-4119-b513-2e0ede6e8e8a-kube-api-access-bnp87\") pod \"olm-operator-6b444d44fb-2c4tv\" (UID: \"38a0c15f-4422-4119-b513-2e0ede6e8e8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157440 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k2qw\" (UniqueName: \"kubernetes.io/projected/a5cdccf3-8273-4553-abbf-fc0bedcf73fc-kube-api-access-8k2qw\") pod \"ingress-canary-zp567\" (UID: \"a5cdccf3-8273-4553-abbf-fc0bedcf73fc\") " pod="openshift-ingress-canary/ingress-canary-zp567" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.157457 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f43cb7f-513b-4958-88f6-e43a01c7adda-trusted-ca\") pod \"ingress-operator-5b745b69d9-mgzl9\" (UID: \"9f43cb7f-513b-4958-88f6-e43a01c7adda\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.158526 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/233ad320-ca2f-40bf-bdcf-64e5aed5509f-trusted-ca\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.159481 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-audit-policies\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.160024 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d10811c5-1f44-4668-a5f8-9690f2d81983-etcd-service-ca\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.160189 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj"] Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.160517 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.160716 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/38a0c15f-4422-4119-b513-2e0ede6e8e8a-srv-cert\") pod \"olm-operator-6b444d44fb-2c4tv\" (UID: \"38a0c15f-4422-4119-b513-2e0ede6e8e8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.160855 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.161510 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/000c6858-a809-46e4-bdcb-b150477e5ce2-audit-dir\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.161542 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/920cdae1-9503-4181-92a2-d62969908d1d-images\") pod \"machine-config-operator-74547568cd-9dhq6\" (UID: \"920cdae1-9503-4181-92a2-d62969908d1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.161558 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9ff0349-32de-4d48-87ad-3e206ec56b2e-service-ca-bundle\") pod \"authentication-operator-69f744f599-6qzsg\" (UID: \"e9ff0349-32de-4d48-87ad-3e206ec56b2e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.161681 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/000c6858-a809-46e4-bdcb-b150477e5ce2-audit-dir\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.162293 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/861e4822-a809-46c7-af83-06e8978ca8aa-service-ca-bundle\") pod \"router-default-5444994796-mmxrh\" (UID: \"861e4822-a809-46c7-af83-06e8978ca8aa\") " pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.163001 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/0647d456-31a3-4906-9d62-f43b4cfef5cb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dgkrj\" (UID: \"0647d456-31a3-4906-9d62-f43b4cfef5cb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dgkrj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.163011 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/79fbf847-050b-404d-8e9a-7f16f42089dc-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-62ppz\" (UID: \"79fbf847-050b-404d-8e9a-7f16f42089dc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.163075 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.163255 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/957c4786-76f4-4437-b56a-d42dee2c65b5-node-bootstrap-token\") pod \"machine-config-server-tclrk\" (UID: \"957c4786-76f4-4437-b56a-d42dee2c65b5\") " pod="openshift-machine-config-operator/machine-config-server-tclrk" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.163361 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/18bfd488-170c-457a-9abb-7b816f932bd0-srv-cert\") pod \"catalog-operator-68c6474976-6rqtd\" (UID: \"18bfd488-170c-457a-9abb-7b816f932bd0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.163479 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.163600 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/861e4822-a809-46c7-af83-06e8978ca8aa-stats-auth\") pod \"router-default-5444994796-mmxrh\" (UID: \"861e4822-a809-46c7-af83-06e8978ca8aa\") " pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.163737 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b5ee7463-4c14-4b15-974d-363744e4aaeb-webhook-cert\") pod \"packageserver-d55dfcdfc-nn8tj\" (UID: \"b5ee7463-4c14-4b15-974d-363744e4aaeb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.163838 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47f6r\" (UniqueName: \"kubernetes.io/projected/08021692-e26d-42ab-85a2-136b1bd2b4ed-kube-api-access-47f6r\") pod \"machine-config-controller-84d6567774-wbb9n\" (UID: \"08021692-e26d-42ab-85a2-136b1bd2b4ed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.163935 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktx2b\" (UniqueName: \"kubernetes.io/projected/e98c9985-31a9-4f14-818d-4519f0a3131b-kube-api-access-ktx2b\") pod \"service-ca-9c57cc56f-tqxb8\" (UID: \"e98c9985-31a9-4f14-818d-4519f0a3131b\") " pod="openshift-service-ca/service-ca-9c57cc56f-tqxb8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.164148 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.165514 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9f43cb7f-513b-4958-88f6-e43a01c7adda-metrics-tls\") pod \"ingress-operator-5b745b69d9-mgzl9\" (UID: \"9f43cb7f-513b-4958-88f6-e43a01c7adda\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.165543 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.165574 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b5ee7463-4c14-4b15-974d-363744e4aaeb-tmpfs\") pod \"packageserver-d55dfcdfc-nn8tj\" (UID: \"b5ee7463-4c14-4b15-974d-363744e4aaeb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.165604 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1905e8ad-dfcf-4ce5-9565-28562f0832bc-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2fq8f\" (UID: \"1905e8ad-dfcf-4ce5-9565-28562f0832bc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.165877 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cdcade7f-c50d-4690-a339-db8dcc9a7ef0-config-volume\") pod \"dns-default-f47vx\" (UID: \"cdcade7f-c50d-4690-a339-db8dcc9a7ef0\") " pod="openshift-dns/dns-default-f47vx" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.167492 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.168195 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/861e4822-a809-46c7-af83-06e8978ca8aa-metrics-certs\") pod \"router-default-5444994796-mmxrh\" (UID: \"861e4822-a809-46c7-af83-06e8978ca8aa\") " pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.168260 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqxmk\" (UniqueName: \"kubernetes.io/projected/e9ff0349-32de-4d48-87ad-3e206ec56b2e-kube-api-access-nqxmk\") pod \"authentication-operator-69f744f599-6qzsg\" (UID: \"e9ff0349-32de-4d48-87ad-3e206ec56b2e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.168285 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.168329 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/18bfd488-170c-457a-9abb-7b816f932bd0-profile-collector-cert\") pod \"catalog-operator-68c6474976-6rqtd\" (UID: \"18bfd488-170c-457a-9abb-7b816f932bd0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.168353 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb71fca6-a96f-4438-84da-e1a6c7c58027-config\") pod \"service-ca-operator-777779d784-7hgq8\" (UID: \"fb71fca6-a96f-4438-84da-e1a6c7c58027\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.168771 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f43cb7f-513b-4958-88f6-e43a01c7adda-trusted-ca\") pod \"ingress-operator-5b745b69d9-mgzl9\" (UID: \"9f43cb7f-513b-4958-88f6-e43a01c7adda\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.169291 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-mountpoint-dir\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.169369 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c85a9a8a-89cf-4c60-9619-879f64d40195-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsqzh\" (UID: \"c85a9a8a-89cf-4c60-9619-879f64d40195\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.169406 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a3f377d2-10db-41db-9361-06abff2e0d29-console-serving-cert\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.169449 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/233ad320-ca2f-40bf-bdcf-64e5aed5509f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.169825 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-62l7r"] Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.170023 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c85a9a8a-89cf-4c60-9619-879f64d40195-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsqzh\" (UID: \"c85a9a8a-89cf-4c60-9619-879f64d40195\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.170065 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.170241 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/233ad320-ca2f-40bf-bdcf-64e5aed5509f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.170412 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9ff0349-32de-4d48-87ad-3e206ec56b2e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6qzsg\" (UID: \"e9ff0349-32de-4d48-87ad-3e206ec56b2e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.170474 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4e4767a8-ff51-43c7-94cc-625c9b8689a3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mzgjp\" (UID: \"4e4767a8-ff51-43c7-94cc-625c9b8689a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.170637 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70d37d83-4fab-49db-8cb5-b98f50c02a0e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-db5db\" (UID: \"70d37d83-4fab-49db-8cb5-b98f50c02a0e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.170671 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e-config\") pod \"kube-apiserver-operator-766d6c64bb-djtqn\" (UID: \"7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.170692 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5fa998f-665b-421d-b589-e2531df3b2c2-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cvm28\" (UID: \"f5fa998f-665b-421d-b589-e2531df3b2c2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.170723 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d10811c5-1f44-4668-a5f8-9690f2d81983-etcd-client\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.170740 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/861e4822-a809-46c7-af83-06e8978ca8aa-default-certificate\") pod \"router-default-5444994796-mmxrh\" (UID: \"861e4822-a809-46c7-af83-06e8978ca8aa\") " pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.170774 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm24h\" (UniqueName: \"kubernetes.io/projected/4fbac3c5-6a92-4cc5-980d-5d66576df0e8-kube-api-access-fm24h\") pod \"multus-admission-controller-857f4d67dd-vrgcp\" (UID: \"4fbac3c5-6a92-4cc5-980d-5d66576df0e8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vrgcp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.170790 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a5cdccf3-8273-4553-abbf-fc0bedcf73fc-cert\") pod \"ingress-canary-zp567\" (UID: \"a5cdccf3-8273-4553-abbf-fc0bedcf73fc\") " pod="openshift-ingress-canary/ingress-canary-zp567" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.172130 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70d37d83-4fab-49db-8cb5-b98f50c02a0e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-db5db\" (UID: \"70d37d83-4fab-49db-8cb5-b98f50c02a0e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.172643 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9ff0349-32de-4d48-87ad-3e206ec56b2e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6qzsg\" (UID: \"e9ff0349-32de-4d48-87ad-3e206ec56b2e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.172757 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d10811c5-1f44-4668-a5f8-9690f2d81983-serving-cert\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.173217 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/861e4822-a809-46c7-af83-06e8978ca8aa-stats-auth\") pod \"router-default-5444994796-mmxrh\" (UID: \"861e4822-a809-46c7-af83-06e8978ca8aa\") " pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.173427 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.173551 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/33ec7e32-c944-4b4c-80c8-4f9a2ad49168-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5r9n4\" (UID: \"33ec7e32-c944-4b4c-80c8-4f9a2ad49168\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.173675 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.173809 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b5ee7463-4c14-4b15-974d-363744e4aaeb-apiservice-cert\") pod \"packageserver-d55dfcdfc-nn8tj\" (UID: \"b5ee7463-4c14-4b15-974d-363744e4aaeb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.173936 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3b036b1c-d79d-49d5-89ab-c669a5f8f47b-metrics-tls\") pod \"dns-operator-744455d44c-ln6xb\" (UID: \"3b036b1c-d79d-49d5-89ab-c669a5f8f47b\") " pod="openshift-dns-operator/dns-operator-744455d44c-ln6xb" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.174030 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c85a9a8a-89cf-4c60-9619-879f64d40195-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsqzh\" (UID: \"c85a9a8a-89cf-4c60-9619-879f64d40195\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.174068 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e-config\") pod \"kube-apiserver-operator-766d6c64bb-djtqn\" (UID: \"7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.174216 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slmtl\" (UniqueName: \"kubernetes.io/projected/18bfd488-170c-457a-9abb-7b816f932bd0-kube-api-access-slmtl\") pod \"catalog-operator-68c6474976-6rqtd\" (UID: \"18bfd488-170c-457a-9abb-7b816f932bd0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.174297 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsmqk\" (UniqueName: \"kubernetes.io/projected/b5ee7463-4c14-4b15-974d-363744e4aaeb-kube-api-access-lsmqk\") pod \"packageserver-d55dfcdfc-nn8tj\" (UID: \"b5ee7463-4c14-4b15-974d-363744e4aaeb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.174782 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70d37d83-4fab-49db-8cb5-b98f50c02a0e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-db5db\" (UID: \"70d37d83-4fab-49db-8cb5-b98f50c02a0e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.174869 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkb2v\" (UniqueName: \"kubernetes.io/projected/c85a9a8a-89cf-4c60-9619-879f64d40195-kube-api-access-rkb2v\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsqzh\" (UID: \"c85a9a8a-89cf-4c60-9619-879f64d40195\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.174895 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh5dl\" (UniqueName: \"kubernetes.io/projected/33ec7e32-c944-4b4c-80c8-4f9a2ad49168-kube-api-access-qh5dl\") pod \"package-server-manager-789f6589d5-5r9n4\" (UID: \"33ec7e32-c944-4b4c-80c8-4f9a2ad49168\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.174915 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a3f377d2-10db-41db-9361-06abff2e0d29-console-oauth-config\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175121 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d10811c5-1f44-4668-a5f8-9690f2d81983-config\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175166 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d10811c5-1f44-4668-a5f8-9690f2d81983-etcd-ca\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175191 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmv2r\" (UniqueName: \"kubernetes.io/projected/861e4822-a809-46c7-af83-06e8978ca8aa-kube-api-access-qmv2r\") pod \"router-default-5444994796-mmxrh\" (UID: \"861e4822-a809-46c7-af83-06e8978ca8aa\") " pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175240 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mmqr\" (UniqueName: \"kubernetes.io/projected/3b036b1c-d79d-49d5-89ab-c669a5f8f47b-kube-api-access-7mmqr\") pod \"dns-operator-744455d44c-ln6xb\" (UID: \"3b036b1c-d79d-49d5-89ab-c669a5f8f47b\") " pod="openshift-dns-operator/dns-operator-744455d44c-ln6xb" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175269 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx9b8\" (UniqueName: \"kubernetes.io/projected/fb71fca6-a96f-4438-84da-e1a6c7c58027-kube-api-access-qx9b8\") pod \"service-ca-operator-777779d784-7hgq8\" (UID: \"fb71fca6-a96f-4438-84da-e1a6c7c58027\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175317 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9ff0349-32de-4d48-87ad-3e206ec56b2e-serving-cert\") pod \"authentication-operator-69f744f599-6qzsg\" (UID: \"e9ff0349-32de-4d48-87ad-3e206ec56b2e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175335 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/861e4822-a809-46c7-af83-06e8978ca8aa-default-certificate\") pod \"router-default-5444994796-mmxrh\" (UID: \"861e4822-a809-46c7-af83-06e8978ca8aa\") " pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175349 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-djtqn\" (UID: \"7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175371 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-trusted-ca-bundle\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175406 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dfts\" (UniqueName: \"kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-kube-api-access-4dfts\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175430 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb71fca6-a96f-4438-84da-e1a6c7c58027-serving-cert\") pod \"service-ca-operator-777779d784-7hgq8\" (UID: \"fb71fca6-a96f-4438-84da-e1a6c7c58027\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175456 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/233ad320-ca2f-40bf-bdcf-64e5aed5509f-registry-certificates\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175478 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4e4767a8-ff51-43c7-94cc-625c9b8689a3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mzgjp\" (UID: \"4e4767a8-ff51-43c7-94cc-625c9b8689a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175518 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/38a0c15f-4422-4119-b513-2e0ede6e8e8a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2c4tv\" (UID: \"38a0c15f-4422-4119-b513-2e0ede6e8e8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175543 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4f1c998-fb99-4827-a353-d8b59a309cee-config-volume\") pod \"collect-profiles-29413665-w4p5k\" (UID: \"b4f1c998-fb99-4827-a353-d8b59a309cee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175592 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175618 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f43cb7f-513b-4958-88f6-e43a01c7adda-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mgzl9\" (UID: \"9f43cb7f-513b-4958-88f6-e43a01c7adda\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175638 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-console-config\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175677 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4f1c998-fb99-4827-a353-d8b59a309cee-secret-volume\") pod \"collect-profiles-29413665-w4p5k\" (UID: \"b4f1c998-fb99-4827-a353-d8b59a309cee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175705 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-registry-tls\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175746 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4fbac3c5-6a92-4cc5-980d-5d66576df0e8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vrgcp\" (UID: \"4fbac3c5-6a92-4cc5-980d-5d66576df0e8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vrgcp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175766 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-oauth-serving-cert\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175789 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/08021692-e26d-42ab-85a2-136b1bd2b4ed-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wbb9n\" (UID: \"08021692-e26d-42ab-85a2-136b1bd2b4ed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175811 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvsr6\" (UniqueName: \"kubernetes.io/projected/b4f1c998-fb99-4827-a353-d8b59a309cee-kube-api-access-kvsr6\") pod \"collect-profiles-29413665-w4p5k\" (UID: \"b4f1c998-fb99-4827-a353-d8b59a309cee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175837 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/920cdae1-9503-4181-92a2-d62969908d1d-proxy-tls\") pod \"machine-config-operator-74547568cd-9dhq6\" (UID: \"920cdae1-9503-4181-92a2-d62969908d1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175857 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d10811c5-1f44-4668-a5f8-9690f2d81983-config\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.175860 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5fa998f-665b-421d-b589-e2531df3b2c2-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cvm28\" (UID: \"f5fa998f-665b-421d-b589-e2531df3b2c2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.176243 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d10811c5-1f44-4668-a5f8-9690f2d81983-etcd-ca\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.176335 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-djtqn\" (UID: \"7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.176640 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d10811c5-1f44-4668-a5f8-9690f2d81983-serving-cert\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.177512 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.177783 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/233ad320-ca2f-40bf-bdcf-64e5aed5509f-registry-certificates\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.178194 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c85a9a8a-89cf-4c60-9619-879f64d40195-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsqzh\" (UID: \"c85a9a8a-89cf-4c60-9619-879f64d40195\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.178214 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.180307 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/38a0c15f-4422-4119-b513-2e0ede6e8e8a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2c4tv\" (UID: \"38a0c15f-4422-4119-b513-2e0ede6e8e8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.181336 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.181788 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d10811c5-1f44-4668-a5f8-9690f2d81983-etcd-client\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.182044 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9ff0349-32de-4d48-87ad-3e206ec56b2e-serving-cert\") pod \"authentication-operator-69f744f599-6qzsg\" (UID: \"e9ff0349-32de-4d48-87ad-3e206ec56b2e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.182149 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.183350 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/861e4822-a809-46c7-af83-06e8978ca8aa-metrics-certs\") pod \"router-default-5444994796-mmxrh\" (UID: \"861e4822-a809-46c7-af83-06e8978ca8aa\") " pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.184222 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3b036b1c-d79d-49d5-89ab-c669a5f8f47b-metrics-tls\") pod \"dns-operator-744455d44c-ln6xb\" (UID: \"3b036b1c-d79d-49d5-89ab-c669a5f8f47b\") " pod="openshift-dns-operator/dns-operator-744455d44c-ln6xb" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.185145 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-registry-tls\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.205572 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/79fbf847-050b-404d-8e9a-7f16f42089dc-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-62ppz\" (UID: \"79fbf847-050b-404d-8e9a-7f16f42089dc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.223576 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd2nl\" (UniqueName: \"kubernetes.io/projected/79fbf847-050b-404d-8e9a-7f16f42089dc-kube-api-access-qd2nl\") pod \"cluster-image-registry-operator-dc59b4c8b-62ppz\" (UID: \"79fbf847-050b-404d-8e9a-7f16f42089dc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.262651 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmlrq\" (UniqueName: \"kubernetes.io/projected/d10811c5-1f44-4668-a5f8-9690f2d81983-kube-api-access-gmlrq\") pod \"etcd-operator-b45778765-fd998\" (UID: \"d10811c5-1f44-4668-a5f8-9690f2d81983\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.277156 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/18bfd488-170c-457a-9abb-7b816f932bd0-profile-collector-cert\") pod \"catalog-operator-68c6474976-6rqtd\" (UID: \"18bfd488-170c-457a-9abb-7b816f932bd0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.277198 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb71fca6-a96f-4438-84da-e1a6c7c58027-config\") pod \"service-ca-operator-777779d784-7hgq8\" (UID: \"fb71fca6-a96f-4438-84da-e1a6c7c58027\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.277224 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-mountpoint-dir\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.277422 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-mountpoint-dir\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.277816 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a3f377d2-10db-41db-9361-06abff2e0d29-console-serving-cert\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279235 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4e4767a8-ff51-43c7-94cc-625c9b8689a3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mzgjp\" (UID: \"4e4767a8-ff51-43c7-94cc-625c9b8689a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279265 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5fa998f-665b-421d-b589-e2531df3b2c2-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cvm28\" (UID: \"f5fa998f-665b-421d-b589-e2531df3b2c2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.278092 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb71fca6-a96f-4438-84da-e1a6c7c58027-config\") pod \"service-ca-operator-777779d784-7hgq8\" (UID: \"fb71fca6-a96f-4438-84da-e1a6c7c58027\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279292 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm24h\" (UniqueName: \"kubernetes.io/projected/4fbac3c5-6a92-4cc5-980d-5d66576df0e8-kube-api-access-fm24h\") pod \"multus-admission-controller-857f4d67dd-vrgcp\" (UID: \"4fbac3c5-6a92-4cc5-980d-5d66576df0e8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vrgcp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279405 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a5cdccf3-8273-4553-abbf-fc0bedcf73fc-cert\") pod \"ingress-canary-zp567\" (UID: \"a5cdccf3-8273-4553-abbf-fc0bedcf73fc\") " pod="openshift-ingress-canary/ingress-canary-zp567" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279443 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/33ec7e32-c944-4b4c-80c8-4f9a2ad49168-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5r9n4\" (UID: \"33ec7e32-c944-4b4c-80c8-4f9a2ad49168\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279461 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b5ee7463-4c14-4b15-974d-363744e4aaeb-apiservice-cert\") pod \"packageserver-d55dfcdfc-nn8tj\" (UID: \"b5ee7463-4c14-4b15-974d-363744e4aaeb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279483 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slmtl\" (UniqueName: \"kubernetes.io/projected/18bfd488-170c-457a-9abb-7b816f932bd0-kube-api-access-slmtl\") pod \"catalog-operator-68c6474976-6rqtd\" (UID: \"18bfd488-170c-457a-9abb-7b816f932bd0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279502 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsmqk\" (UniqueName: \"kubernetes.io/projected/b5ee7463-4c14-4b15-974d-363744e4aaeb-kube-api-access-lsmqk\") pod \"packageserver-d55dfcdfc-nn8tj\" (UID: \"b5ee7463-4c14-4b15-974d-363744e4aaeb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279530 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh5dl\" (UniqueName: \"kubernetes.io/projected/33ec7e32-c944-4b4c-80c8-4f9a2ad49168-kube-api-access-qh5dl\") pod \"package-server-manager-789f6589d5-5r9n4\" (UID: \"33ec7e32-c944-4b4c-80c8-4f9a2ad49168\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279573 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a3f377d2-10db-41db-9361-06abff2e0d29-console-oauth-config\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279591 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx9b8\" (UniqueName: \"kubernetes.io/projected/fb71fca6-a96f-4438-84da-e1a6c7c58027-kube-api-access-qx9b8\") pod \"service-ca-operator-777779d784-7hgq8\" (UID: \"fb71fca6-a96f-4438-84da-e1a6c7c58027\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279630 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-trusted-ca-bundle\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279659 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb71fca6-a96f-4438-84da-e1a6c7c58027-serving-cert\") pod \"service-ca-operator-777779d784-7hgq8\" (UID: \"fb71fca6-a96f-4438-84da-e1a6c7c58027\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279686 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4e4767a8-ff51-43c7-94cc-625c9b8689a3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mzgjp\" (UID: \"4e4767a8-ff51-43c7-94cc-625c9b8689a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279713 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4f1c998-fb99-4827-a353-d8b59a309cee-config-volume\") pod \"collect-profiles-29413665-w4p5k\" (UID: \"b4f1c998-fb99-4827-a353-d8b59a309cee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279741 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-console-config\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279782 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4f1c998-fb99-4827-a353-d8b59a309cee-secret-volume\") pod \"collect-profiles-29413665-w4p5k\" (UID: \"b4f1c998-fb99-4827-a353-d8b59a309cee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279798 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvsr6\" (UniqueName: \"kubernetes.io/projected/b4f1c998-fb99-4827-a353-d8b59a309cee-kube-api-access-kvsr6\") pod \"collect-profiles-29413665-w4p5k\" (UID: \"b4f1c998-fb99-4827-a353-d8b59a309cee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279816 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4fbac3c5-6a92-4cc5-980d-5d66576df0e8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vrgcp\" (UID: \"4fbac3c5-6a92-4cc5-980d-5d66576df0e8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vrgcp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279833 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-oauth-serving-cert\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279849 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/08021692-e26d-42ab-85a2-136b1bd2b4ed-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wbb9n\" (UID: \"08021692-e26d-42ab-85a2-136b1bd2b4ed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279875 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5fa998f-665b-421d-b589-e2531df3b2c2-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cvm28\" (UID: \"f5fa998f-665b-421d-b589-e2531df3b2c2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279901 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1905e8ad-dfcf-4ce5-9565-28562f0832bc-config\") pod \"kube-controller-manager-operator-78b949d7b-2fq8f\" (UID: \"1905e8ad-dfcf-4ce5-9565-28562f0832bc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279935 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-csi-data-dir\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279961 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv6mt\" (UniqueName: \"kubernetes.io/projected/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-kube-api-access-gv6mt\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.279981 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e98c9985-31a9-4f14-818d-4519f0a3131b-signing-key\") pod \"service-ca-9c57cc56f-tqxb8\" (UID: \"e98c9985-31a9-4f14-818d-4519f0a3131b\") " pod="openshift-service-ca/service-ca-9c57cc56f-tqxb8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280005 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-plugins-dir\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280022 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-socket-dir\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280049 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-service-ca\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280066 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5fa998f-665b-421d-b589-e2531df3b2c2-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cvm28\" (UID: \"f5fa998f-665b-421d-b589-e2531df3b2c2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280082 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/957c4786-76f4-4437-b56a-d42dee2c65b5-certs\") pod \"machine-config-server-tclrk\" (UID: \"957c4786-76f4-4437-b56a-d42dee2c65b5\") " pod="openshift-machine-config-operator/machine-config-server-tclrk" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280096 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/08021692-e26d-42ab-85a2-136b1bd2b4ed-proxy-tls\") pod \"machine-config-controller-84d6567774-wbb9n\" (UID: \"08021692-e26d-42ab-85a2-136b1bd2b4ed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280115 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-registration-dir\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280153 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280191 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cdcade7f-c50d-4690-a339-db8dcc9a7ef0-metrics-tls\") pod \"dns-default-f47vx\" (UID: \"cdcade7f-c50d-4690-a339-db8dcc9a7ef0\") " pod="openshift-dns/dns-default-f47vx" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280208 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5m5h\" (UniqueName: \"kubernetes.io/projected/4e4767a8-ff51-43c7-94cc-625c9b8689a3-kube-api-access-j5m5h\") pod \"marketplace-operator-79b997595-mzgjp\" (UID: \"4e4767a8-ff51-43c7-94cc-625c9b8689a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280242 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc5f8\" (UniqueName: \"kubernetes.io/projected/957c4786-76f4-4437-b56a-d42dee2c65b5-kube-api-access-tc5f8\") pod \"machine-config-server-tclrk\" (UID: \"957c4786-76f4-4437-b56a-d42dee2c65b5\") " pod="openshift-machine-config-operator/machine-config-server-tclrk" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280273 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcvgf\" (UniqueName: \"kubernetes.io/projected/a3f377d2-10db-41db-9361-06abff2e0d29-kube-api-access-lcvgf\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280307 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzjbm\" (UniqueName: \"kubernetes.io/projected/2f20b247-6932-459d-a19e-fd200ba45136-kube-api-access-zzjbm\") pod \"migrator-59844c95c7-bl2l4\" (UID: \"2f20b247-6932-459d-a19e-fd200ba45136\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bl2l4" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280329 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gll4j\" (UniqueName: \"kubernetes.io/projected/cdcade7f-c50d-4690-a339-db8dcc9a7ef0-kube-api-access-gll4j\") pod \"dns-default-f47vx\" (UID: \"cdcade7f-c50d-4690-a339-db8dcc9a7ef0\") " pod="openshift-dns/dns-default-f47vx" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280346 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1905e8ad-dfcf-4ce5-9565-28562f0832bc-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2fq8f\" (UID: \"1905e8ad-dfcf-4ce5-9565-28562f0832bc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280370 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e98c9985-31a9-4f14-818d-4519f0a3131b-signing-cabundle\") pod \"service-ca-9c57cc56f-tqxb8\" (UID: \"e98c9985-31a9-4f14-818d-4519f0a3131b\") " pod="openshift-service-ca/service-ca-9c57cc56f-tqxb8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280399 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k2qw\" (UniqueName: \"kubernetes.io/projected/a5cdccf3-8273-4553-abbf-fc0bedcf73fc-kube-api-access-8k2qw\") pod \"ingress-canary-zp567\" (UID: \"a5cdccf3-8273-4553-abbf-fc0bedcf73fc\") " pod="openshift-ingress-canary/ingress-canary-zp567" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280419 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/957c4786-76f4-4437-b56a-d42dee2c65b5-node-bootstrap-token\") pod \"machine-config-server-tclrk\" (UID: \"957c4786-76f4-4437-b56a-d42dee2c65b5\") " pod="openshift-machine-config-operator/machine-config-server-tclrk" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280434 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/18bfd488-170c-457a-9abb-7b816f932bd0-srv-cert\") pod \"catalog-operator-68c6474976-6rqtd\" (UID: \"18bfd488-170c-457a-9abb-7b816f932bd0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280451 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b5ee7463-4c14-4b15-974d-363744e4aaeb-webhook-cert\") pod \"packageserver-d55dfcdfc-nn8tj\" (UID: \"b5ee7463-4c14-4b15-974d-363744e4aaeb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280466 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47f6r\" (UniqueName: \"kubernetes.io/projected/08021692-e26d-42ab-85a2-136b1bd2b4ed-kube-api-access-47f6r\") pod \"machine-config-controller-84d6567774-wbb9n\" (UID: \"08021692-e26d-42ab-85a2-136b1bd2b4ed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280482 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktx2b\" (UniqueName: \"kubernetes.io/projected/e98c9985-31a9-4f14-818d-4519f0a3131b-kube-api-access-ktx2b\") pod \"service-ca-9c57cc56f-tqxb8\" (UID: \"e98c9985-31a9-4f14-818d-4519f0a3131b\") " pod="openshift-service-ca/service-ca-9c57cc56f-tqxb8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280501 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b5ee7463-4c14-4b15-974d-363744e4aaeb-tmpfs\") pod \"packageserver-d55dfcdfc-nn8tj\" (UID: \"b5ee7463-4c14-4b15-974d-363744e4aaeb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280520 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1905e8ad-dfcf-4ce5-9565-28562f0832bc-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2fq8f\" (UID: \"1905e8ad-dfcf-4ce5-9565-28562f0832bc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.280538 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cdcade7f-c50d-4690-a339-db8dcc9a7ef0-config-volume\") pod \"dns-default-f47vx\" (UID: \"cdcade7f-c50d-4690-a339-db8dcc9a7ef0\") " pod="openshift-dns/dns-default-f47vx" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.281353 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4e4767a8-ff51-43c7-94cc-625c9b8689a3-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mzgjp\" (UID: \"4e4767a8-ff51-43c7-94cc-625c9b8689a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.281971 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-console-config\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.282679 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-trusted-ca-bundle\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.283485 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4f1c998-fb99-4827-a353-d8b59a309cee-config-volume\") pod \"collect-profiles-29413665-w4p5k\" (UID: \"b4f1c998-fb99-4827-a353-d8b59a309cee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.283753 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-oauth-serving-cert\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.283908 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a3f377d2-10db-41db-9361-06abff2e0d29-console-oauth-config\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.284274 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/18bfd488-170c-457a-9abb-7b816f932bd0-profile-collector-cert\") pod \"catalog-operator-68c6474976-6rqtd\" (UID: \"18bfd488-170c-457a-9abb-7b816f932bd0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.284406 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/08021692-e26d-42ab-85a2-136b1bd2b4ed-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wbb9n\" (UID: \"08021692-e26d-42ab-85a2-136b1bd2b4ed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.284443 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-csi-data-dir\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.284536 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b5ee7463-4c14-4b15-974d-363744e4aaeb-apiservice-cert\") pod \"packageserver-d55dfcdfc-nn8tj\" (UID: \"b5ee7463-4c14-4b15-974d-363744e4aaeb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.284553 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-socket-dir\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.284617 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-plugins-dir\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.285122 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-service-ca\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.285661 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb71fca6-a96f-4438-84da-e1a6c7c58027-serving-cert\") pod \"service-ca-operator-777779d784-7hgq8\" (UID: \"fb71fca6-a96f-4438-84da-e1a6c7c58027\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.285916 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-registration-dir\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.286125 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5fa998f-665b-421d-b589-e2531df3b2c2-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cvm28\" (UID: \"f5fa998f-665b-421d-b589-e2531df3b2c2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28" Dec 04 03:57:02 crc kubenswrapper[4806]: E1204 03:57:02.286527 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:02.786510613 +0000 UTC m=+137.645023541 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.287248 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b5ee7463-4c14-4b15-974d-363744e4aaeb-tmpfs\") pod \"packageserver-d55dfcdfc-nn8tj\" (UID: \"b5ee7463-4c14-4b15-974d-363744e4aaeb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.287889 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1905e8ad-dfcf-4ce5-9565-28562f0832bc-config\") pod \"kube-controller-manager-operator-78b949d7b-2fq8f\" (UID: \"1905e8ad-dfcf-4ce5-9565-28562f0832bc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.287941 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1905e8ad-dfcf-4ce5-9565-28562f0832bc-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2fq8f\" (UID: \"1905e8ad-dfcf-4ce5-9565-28562f0832bc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.288348 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a5cdccf3-8273-4553-abbf-fc0bedcf73fc-cert\") pod \"ingress-canary-zp567\" (UID: \"a5cdccf3-8273-4553-abbf-fc0bedcf73fc\") " pod="openshift-ingress-canary/ingress-canary-zp567" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.289510 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e98c9985-31a9-4f14-818d-4519f0a3131b-signing-key\") pod \"service-ca-9c57cc56f-tqxb8\" (UID: \"e98c9985-31a9-4f14-818d-4519f0a3131b\") " pod="openshift-service-ca/service-ca-9c57cc56f-tqxb8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.289997 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/08021692-e26d-42ab-85a2-136b1bd2b4ed-proxy-tls\") pod \"machine-config-controller-84d6567774-wbb9n\" (UID: \"08021692-e26d-42ab-85a2-136b1bd2b4ed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.290260 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/957c4786-76f4-4437-b56a-d42dee2c65b5-certs\") pod \"machine-config-server-tclrk\" (UID: \"957c4786-76f4-4437-b56a-d42dee2c65b5\") " pod="openshift-machine-config-operator/machine-config-server-tclrk" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.290372 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b5ee7463-4c14-4b15-974d-363744e4aaeb-webhook-cert\") pod \"packageserver-d55dfcdfc-nn8tj\" (UID: \"b5ee7463-4c14-4b15-974d-363744e4aaeb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.292292 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5fa998f-665b-421d-b589-e2531df3b2c2-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cvm28\" (UID: \"f5fa998f-665b-421d-b589-e2531df3b2c2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.292562 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4fbac3c5-6a92-4cc5-980d-5d66576df0e8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vrgcp\" (UID: \"4fbac3c5-6a92-4cc5-980d-5d66576df0e8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vrgcp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.292654 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4f1c998-fb99-4827-a353-d8b59a309cee-secret-volume\") pod \"collect-profiles-29413665-w4p5k\" (UID: \"b4f1c998-fb99-4827-a353-d8b59a309cee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.293389 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5jq4\" (UniqueName: \"kubernetes.io/projected/46000442-50bc-4480-9939-6a56032d3d2d-kube-api-access-w5jq4\") pod \"downloads-7954f5f757-8r5hx\" (UID: \"46000442-50bc-4480-9939-6a56032d3d2d\") " pod="openshift-console/downloads-7954f5f757-8r5hx" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.293810 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/33ec7e32-c944-4b4c-80c8-4f9a2ad49168-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5r9n4\" (UID: \"33ec7e32-c944-4b4c-80c8-4f9a2ad49168\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.294201 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/957c4786-76f4-4437-b56a-d42dee2c65b5-node-bootstrap-token\") pod \"machine-config-server-tclrk\" (UID: \"957c4786-76f4-4437-b56a-d42dee2c65b5\") " pod="openshift-machine-config-operator/machine-config-server-tclrk" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.295056 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a3f377d2-10db-41db-9361-06abff2e0d29-console-serving-cert\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.296232 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/18bfd488-170c-457a-9abb-7b816f932bd0-srv-cert\") pod \"catalog-operator-68c6474976-6rqtd\" (UID: \"18bfd488-170c-457a-9abb-7b816f932bd0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.296657 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4e4767a8-ff51-43c7-94cc-625c9b8689a3-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mzgjp\" (UID: \"4e4767a8-ff51-43c7-94cc-625c9b8689a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.304244 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnp87\" (UniqueName: \"kubernetes.io/projected/38a0c15f-4422-4119-b513-2e0ede6e8e8a-kube-api-access-bnp87\") pod \"olm-operator-6b444d44fb-2c4tv\" (UID: \"38a0c15f-4422-4119-b513-2e0ede6e8e8a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.320877 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-bound-sa-token\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.360678 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmfkm\" (UniqueName: \"kubernetes.io/projected/0647d456-31a3-4906-9d62-f43b4cfef5cb-kube-api-access-mmfkm\") pod \"control-plane-machine-set-operator-78cbb6b69f-dgkrj\" (UID: \"0647d456-31a3-4906-9d62-f43b4cfef5cb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dgkrj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.381413 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:02 crc kubenswrapper[4806]: E1204 03:57:02.381556 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:02.881524426 +0000 UTC m=+137.740037364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.381816 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: E1204 03:57:02.382184 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:02.882172848 +0000 UTC m=+137.740685776 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.382432 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxnc5\" (UniqueName: \"kubernetes.io/projected/70d37d83-4fab-49db-8cb5-b98f50c02a0e-kube-api-access-vxnc5\") pod \"openshift-controller-manager-operator-756b6f6bc6-db5db\" (UID: \"70d37d83-4fab-49db-8cb5-b98f50c02a0e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.396780 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8r5hx" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.403635 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqxmk\" (UniqueName: \"kubernetes.io/projected/e9ff0349-32de-4d48-87ad-3e206ec56b2e-kube-api-access-nqxmk\") pod \"authentication-operator-69f744f599-6qzsg\" (UID: \"e9ff0349-32de-4d48-87ad-3e206ec56b2e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.404363 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.411684 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.434165 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.441086 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mmqr\" (UniqueName: \"kubernetes.io/projected/3b036b1c-d79d-49d5-89ab-c669a5f8f47b-kube-api-access-7mmqr\") pod \"dns-operator-744455d44c-ln6xb\" (UID: \"3b036b1c-d79d-49d5-89ab-c669a5f8f47b\") " pod="openshift-dns-operator/dns-operator-744455d44c-ln6xb" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.447971 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.462178 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmv2r\" (UniqueName: \"kubernetes.io/projected/861e4822-a809-46c7-af83-06e8978ca8aa-kube-api-access-qmv2r\") pod \"router-default-5444994796-mmxrh\" (UID: \"861e4822-a809-46c7-af83-06e8978ca8aa\") " pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.463773 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ln6xb" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.482532 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.482911 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dfts\" (UniqueName: \"kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-kube-api-access-4dfts\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: E1204 03:57:02.483080 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:02.983056475 +0000 UTC m=+137.841569593 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.483277 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dgkrj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.505776 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-djtqn\" (UID: \"7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.540973 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cdcade7f-c50d-4690-a339-db8dcc9a7ef0-config-volume\") pod \"dns-default-f47vx\" (UID: \"cdcade7f-c50d-4690-a339-db8dcc9a7ef0\") " pod="openshift-dns/dns-default-f47vx" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.545182 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e98c9985-31a9-4f14-818d-4519f0a3131b-signing-cabundle\") pod \"service-ca-9c57cc56f-tqxb8\" (UID: \"e98c9985-31a9-4f14-818d-4519f0a3131b\") " pod="openshift-service-ca/service-ca-9c57cc56f-tqxb8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.556509 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6c5j\" (UniqueName: \"kubernetes.io/projected/000c6858-a809-46e4-bdcb-b150477e5ce2-kube-api-access-f6c5j\") pod \"oauth-openshift-558db77b4-7m42k\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.556886 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hpkz\" (UniqueName: \"kubernetes.io/projected/920cdae1-9503-4181-92a2-d62969908d1d-kube-api-access-2hpkz\") pod \"machine-config-operator-74547568cd-9dhq6\" (UID: \"920cdae1-9503-4181-92a2-d62969908d1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.561055 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pddk8"] Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.561655 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/233ad320-ca2f-40bf-bdcf-64e5aed5509f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.564822 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cdcade7f-c50d-4690-a339-db8dcc9a7ef0-metrics-tls\") pod \"dns-default-f47vx\" (UID: \"cdcade7f-c50d-4690-a339-db8dcc9a7ef0\") " pod="openshift-dns/dns-default-f47vx" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.564912 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p"] Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.569046 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/920cdae1-9503-4181-92a2-d62969908d1d-proxy-tls\") pod \"machine-config-operator-74547568cd-9dhq6\" (UID: \"920cdae1-9503-4181-92a2-d62969908d1d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" Dec 04 03:57:02 crc kubenswrapper[4806]: W1204 03:57:02.569414 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32ca5dd6_f353_4c80_ad4b_1dacf3057d1d.slice/crio-0bed43133fd1ccfa2758b0f18550d51e8aec4f7009dd924bd6383949f6e821ae WatchSource:0}: Error finding container 0bed43133fd1ccfa2758b0f18550d51e8aec4f7009dd924bd6383949f6e821ae: Status 404 returned error can't find the container with id 0bed43133fd1ccfa2758b0f18550d51e8aec4f7009dd924bd6383949f6e821ae Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.569506 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm24h\" (UniqueName: \"kubernetes.io/projected/4fbac3c5-6a92-4cc5-980d-5d66576df0e8-kube-api-access-fm24h\") pod \"multus-admission-controller-857f4d67dd-vrgcp\" (UID: \"4fbac3c5-6a92-4cc5-980d-5d66576df0e8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vrgcp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.571089 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9f43cb7f-513b-4958-88f6-e43a01c7adda-metrics-tls\") pod \"ingress-operator-5b745b69d9-mgzl9\" (UID: \"9f43cb7f-513b-4958-88f6-e43a01c7adda\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.571786 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75hqw\" (UniqueName: \"kubernetes.io/projected/9f43cb7f-513b-4958-88f6-e43a01c7adda-kube-api-access-75hqw\") pod \"ingress-operator-5b745b69d9-mgzl9\" (UID: \"9f43cb7f-513b-4958-88f6-e43a01c7adda\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.572415 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f43cb7f-513b-4958-88f6-e43a01c7adda-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mgzl9\" (UID: \"9f43cb7f-513b-4958-88f6-e43a01c7adda\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.580653 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh5dl\" (UniqueName: \"kubernetes.io/projected/33ec7e32-c944-4b4c-80c8-4f9a2ad49168-kube-api-access-qh5dl\") pod \"package-server-manager-789f6589d5-5r9n4\" (UID: \"33ec7e32-c944-4b4c-80c8-4f9a2ad49168\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.580653 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml"] Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.582381 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-jhptt"] Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.587725 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkb2v\" (UniqueName: \"kubernetes.io/projected/c85a9a8a-89cf-4c60-9619-879f64d40195-kube-api-access-rkb2v\") pod \"kube-storage-version-migrator-operator-b67b599dd-nsqzh\" (UID: \"c85a9a8a-89cf-4c60-9619-879f64d40195\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.590723 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: E1204 03:57:02.591952 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:03.091874948 +0000 UTC m=+137.950387876 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.593757 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slmtl\" (UniqueName: \"kubernetes.io/projected/18bfd488-170c-457a-9abb-7b816f932bd0-kube-api-access-slmtl\") pod \"catalog-operator-68c6474976-6rqtd\" (UID: \"18bfd488-170c-457a-9abb-7b816f932bd0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" Dec 04 03:57:02 crc kubenswrapper[4806]: W1204 03:57:02.608373 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5761a71_91cf_4781_9d7d_dc3f9b9096d0.slice/crio-6ede1dfb3e13e9e8e1ee86c36c692e07a852bee4113bd114faa9280657569e64 WatchSource:0}: Error finding container 6ede1dfb3e13e9e8e1ee86c36c692e07a852bee4113bd114faa9280657569e64: Status 404 returned error can't find the container with id 6ede1dfb3e13e9e8e1ee86c36c692e07a852bee4113bd114faa9280657569e64 Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.613257 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.639775 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsmqk\" (UniqueName: \"kubernetes.io/projected/b5ee7463-4c14-4b15-974d-363744e4aaeb-kube-api-access-lsmqk\") pod \"packageserver-d55dfcdfc-nn8tj\" (UID: \"b5ee7463-4c14-4b15-974d-363744e4aaeb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.650196 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvsr6\" (UniqueName: \"kubernetes.io/projected/b4f1c998-fb99-4827-a353-d8b59a309cee-kube-api-access-kvsr6\") pod \"collect-profiles-29413665-w4p5k\" (UID: \"b4f1c998-fb99-4827-a353-d8b59a309cee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.681692 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx9b8\" (UniqueName: \"kubernetes.io/projected/fb71fca6-a96f-4438-84da-e1a6c7c58027-kube-api-access-qx9b8\") pod \"service-ca-operator-777779d784-7hgq8\" (UID: \"fb71fca6-a96f-4438-84da-e1a6c7c58027\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.686899 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f5fa998f-665b-421d-b589-e2531df3b2c2-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cvm28\" (UID: \"f5fa998f-665b-421d-b589-e2531df3b2c2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.692817 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.694206 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:02 crc kubenswrapper[4806]: E1204 03:57:02.694753 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:03.194734313 +0000 UTC m=+138.053247241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.708724 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5m5h\" (UniqueName: \"kubernetes.io/projected/4e4767a8-ff51-43c7-94cc-625c9b8689a3-kube-api-access-j5m5h\") pod \"marketplace-operator-79b997595-mzgjp\" (UID: \"4e4767a8-ff51-43c7-94cc-625c9b8689a3\") " pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.719641 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.725023 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc5f8\" (UniqueName: \"kubernetes.io/projected/957c4786-76f4-4437-b56a-d42dee2c65b5-kube-api-access-tc5f8\") pod \"machine-config-server-tclrk\" (UID: \"957c4786-76f4-4437-b56a-d42dee2c65b5\") " pod="openshift-machine-config-operator/machine-config-server-tclrk" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.728179 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcvgf\" (UniqueName: \"kubernetes.io/projected/a3f377d2-10db-41db-9361-06abff2e0d29-kube-api-access-lcvgf\") pod \"console-f9d7485db-z2mdj\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.740636 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.745289 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzjbm\" (UniqueName: \"kubernetes.io/projected/2f20b247-6932-459d-a19e-fd200ba45136-kube-api-access-zzjbm\") pod \"migrator-59844c95c7-bl2l4\" (UID: \"2f20b247-6932-459d-a19e-fd200ba45136\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bl2l4" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.760039 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.762105 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gll4j\" (UniqueName: \"kubernetes.io/projected/cdcade7f-c50d-4690-a339-db8dcc9a7ef0-kube-api-access-gll4j\") pod \"dns-default-f47vx\" (UID: \"cdcade7f-c50d-4690-a339-db8dcc9a7ef0\") " pod="openshift-dns/dns-default-f47vx" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.774091 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.776230 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.783002 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47f6r\" (UniqueName: \"kubernetes.io/projected/08021692-e26d-42ab-85a2-136b1bd2b4ed-kube-api-access-47f6r\") pod \"machine-config-controller-84d6567774-wbb9n\" (UID: \"08021692-e26d-42ab-85a2-136b1bd2b4ed\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.792020 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.795980 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:02 crc kubenswrapper[4806]: E1204 03:57:02.796306 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:03.296290964 +0000 UTC m=+138.154803892 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.798063 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-vrgcp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.811631 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1905e8ad-dfcf-4ce5-9565-28562f0832bc-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2fq8f\" (UID: \"1905e8ad-dfcf-4ce5-9565-28562f0832bc\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.812122 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.824012 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.827161 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.838138 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv6mt\" (UniqueName: \"kubernetes.io/projected/5e02c1ac-06b2-4d89-98f8-545f682c4a4d-kube-api-access-gv6mt\") pod \"csi-hostpathplugin-cj594\" (UID: \"5e02c1ac-06b2-4d89-98f8-545f682c4a4d\") " pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.841455 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.846026 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.850675 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktx2b\" (UniqueName: \"kubernetes.io/projected/e98c9985-31a9-4f14-818d-4519f0a3131b-kube-api-access-ktx2b\") pod \"service-ca-9c57cc56f-tqxb8\" (UID: \"e98c9985-31a9-4f14-818d-4519f0a3131b\") " pod="openshift-service-ca/service-ca-9c57cc56f-tqxb8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.856306 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bl2l4" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.866849 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.871362 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k2qw\" (UniqueName: \"kubernetes.io/projected/a5cdccf3-8273-4553-abbf-fc0bedcf73fc-kube-api-access-8k2qw\") pod \"ingress-canary-zp567\" (UID: \"a5cdccf3-8273-4553-abbf-fc0bedcf73fc\") " pod="openshift-ingress-canary/ingress-canary-zp567" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.880190 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.885275 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.896819 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:02 crc kubenswrapper[4806]: E1204 03:57:02.897223 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:03.397202371 +0000 UTC m=+138.255715299 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.898580 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-tqxb8" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.911417 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-tclrk" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.934653 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-cj594" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.944053 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-zp567" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.953446 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-f47vx" Dec 04 03:57:02 crc kubenswrapper[4806]: I1204 03:57:02.987341 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8r5hx"] Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:02.998726 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:03 crc kubenswrapper[4806]: E1204 03:57:02.999263 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:03.499238168 +0000 UTC m=+138.357751286 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.081692 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" event={"ID":"5229f915-0d18-41e7-abaf-1d2a783670eb","Type":"ContainerStarted","Data":"d3cf3bcf2888c5b489efc4549853df2005ca175152cc538e86bc1b95b2c00068"} Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.093316 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p" event={"ID":"b5761a71-91cf-4781-9d7d-dc3f9b9096d0","Type":"ContainerStarted","Data":"6ede1dfb3e13e9e8e1ee86c36c692e07a852bee4113bd114faa9280657569e64"} Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.104122 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f" Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.104883 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:03 crc kubenswrapper[4806]: E1204 03:57:03.105207 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:03.60517815 +0000 UTC m=+138.463691088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.110598 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:03 crc kubenswrapper[4806]: E1204 03:57:03.111052 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:03.611020393 +0000 UTC m=+138.469533531 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.124336 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" event={"ID":"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8","Type":"ContainerStarted","Data":"47649cc274c5d141b08ded1e10754ce841aec9b3694943991f738cff8e111d60"} Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.139342 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" event={"ID":"7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6","Type":"ContainerStarted","Data":"f06b8acc8bab7f258e08955c38edb63ab2a5d3b9ea398f3b1036f942c6f5167c"} Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.139398 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" event={"ID":"7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6","Type":"ContainerStarted","Data":"3a6c21d0850c699a989f1831d4e3545bee2c86f54f04d91edf6823f8dafdcd84"} Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.156267 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" event={"ID":"ba31bbad-4048-4651-92cd-d0cba1179541","Type":"ContainerStarted","Data":"1f7c7e01328510106a210fb71a27cc64b3ff8cb36df8228ede1cc262141cca52"} Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.181197 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" event={"ID":"40553430-c0eb-45b4-880b-d0e1f85adb33","Type":"ContainerStarted","Data":"5b488f1bef44886558833a859fa9116267ff611d7de3df2b93ffea7bf6c5d52b"} Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.193719 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" event={"ID":"606e54dc-5cc4-4e02-b796-dd37b613a8d0","Type":"ContainerStarted","Data":"000234645f09cc051b8539795308d55d8b15a2c17f9cf79a6a11fd779da39e8b"} Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.217293 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:03 crc kubenswrapper[4806]: E1204 03:57:03.217754 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:03.717735012 +0000 UTC m=+138.576247940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.235160 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv"] Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.259167 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8nlkb" event={"ID":"5b4637a3-8d20-4427-8e71-9b5ab8e6ee8c","Type":"ContainerStarted","Data":"f136dd7e0a085485e8ca7d357763f302cc907daa1b655f3407b2b4d04338f311"} Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.278726 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-62l7r" event={"ID":"32ca5dd6-f353-4c80-ad4b-1dacf3057d1d","Type":"ContainerStarted","Data":"0bed43133fd1ccfa2758b0f18550d51e8aec4f7009dd924bd6383949f6e821ae"} Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.322699 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:03 crc kubenswrapper[4806]: E1204 03:57:03.325065 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:03.825046772 +0000 UTC m=+138.683559700 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.336899 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" event={"ID":"53e674b8-9044-45f1-8aa5-bb81ec63ae4a","Type":"ContainerStarted","Data":"3b925457e44ef277ed6aeeae7191e4a521e364867ae481272124e985896f9b68"} Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.341153 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fd998"] Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.432686 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:03 crc kubenswrapper[4806]: E1204 03:57:03.433391 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:03.933373087 +0000 UTC m=+138.791886015 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.538361 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:03 crc kubenswrapper[4806]: E1204 03:57:03.538950 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:04.038936806 +0000 UTC m=+138.897449734 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.598629 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dgkrj"] Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.599010 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7m42k"] Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.599027 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz"] Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.599042 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ln6xb"] Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.599054 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db"] Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.638838 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:03 crc kubenswrapper[4806]: E1204 03:57:03.639455 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:04.1394356 +0000 UTC m=+138.997948528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.740881 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:03 crc kubenswrapper[4806]: E1204 03:57:03.743850 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:04.243835449 +0000 UTC m=+139.102348377 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:03 crc kubenswrapper[4806]: W1204 03:57:03.775774 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70d37d83_4fab_49db_8cb5_b98f50c02a0e.slice/crio-b12cd11d76785e074d3a2b81957c1659efc480f89209ebd950456702b8008865 WatchSource:0}: Error finding container b12cd11d76785e074d3a2b81957c1659efc480f89209ebd950456702b8008865: Status 404 returned error can't find the container with id b12cd11d76785e074d3a2b81957c1659efc480f89209ebd950456702b8008865 Dec 04 03:57:03 crc kubenswrapper[4806]: W1204 03:57:03.777585 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b036b1c_d79d_49d5_89ab_c669a5f8f47b.slice/crio-c13ff2012a4ce9fc29e03983908cac541c163614d34a52701f073fff338d03dc WatchSource:0}: Error finding container c13ff2012a4ce9fc29e03983908cac541c163614d34a52701f073fff338d03dc: Status 404 returned error can't find the container with id c13ff2012a4ce9fc29e03983908cac541c163614d34a52701f073fff338d03dc Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.833301 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6"] Dec 04 03:57:03 crc kubenswrapper[4806]: W1204 03:57:03.837904 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79fbf847_050b_404d_8e9a_7f16f42089dc.slice/crio-718a22389c7b86ddb3451a7ab0bcea61d00167b7b0569ec5f5ee2ae1b4151894 WatchSource:0}: Error finding container 718a22389c7b86ddb3451a7ab0bcea61d00167b7b0569ec5f5ee2ae1b4151894: Status 404 returned error can't find the container with id 718a22389c7b86ddb3451a7ab0bcea61d00167b7b0569ec5f5ee2ae1b4151894 Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.842513 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:03 crc kubenswrapper[4806]: E1204 03:57:03.842861 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:04.342842221 +0000 UTC m=+139.201355149 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.857352 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6qzsg"] Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.946417 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:03 crc kubenswrapper[4806]: E1204 03:57:03.946810 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:04.446795695 +0000 UTC m=+139.305308623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.987038 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-f47vx"] Dec 04 03:57:03 crc kubenswrapper[4806]: I1204 03:57:03.993997 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k"] Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.048492 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:04 crc kubenswrapper[4806]: E1204 03:57:04.048846 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:04.548815301 +0000 UTC m=+139.407328229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.049168 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:04 crc kubenswrapper[4806]: E1204 03:57:04.049555 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:04.549546036 +0000 UTC m=+139.408058964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:04 crc kubenswrapper[4806]: W1204 03:57:04.070351 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod957c4786_76f4_4437_b56a_d42dee2c65b5.slice/crio-e2015ad621f83db87c029b1ece2d24dd603208120149e24d13c9cc7b6cca6a12 WatchSource:0}: Error finding container e2015ad621f83db87c029b1ece2d24dd603208120149e24d13c9cc7b6cca6a12: Status 404 returned error can't find the container with id e2015ad621f83db87c029b1ece2d24dd603208120149e24d13c9cc7b6cca6a12 Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.150561 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:04 crc kubenswrapper[4806]: E1204 03:57:04.150778 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:04.650758995 +0000 UTC m=+139.509271923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.151288 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:04 crc kubenswrapper[4806]: E1204 03:57:04.151640 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:04.651629264 +0000 UTC m=+139.510142192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:04 crc kubenswrapper[4806]: W1204 03:57:04.162836 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcdcade7f_c50d_4690_a339_db8dcc9a7ef0.slice/crio-4eebc1c3eb7cec23ca89ad4df55f0070504d1a548c3dc0a4bec689dc320115c5 WatchSource:0}: Error finding container 4eebc1c3eb7cec23ca89ad4df55f0070504d1a548c3dc0a4bec689dc320115c5: Status 404 returned error can't find the container with id 4eebc1c3eb7cec23ca89ad4df55f0070504d1a548c3dc0a4bec689dc320115c5 Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.254460 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:04 crc kubenswrapper[4806]: E1204 03:57:04.254783 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:04.75476838 +0000 UTC m=+139.613281298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:04 crc kubenswrapper[4806]: W1204 03:57:04.279407 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4f1c998_fb99_4827_a353_d8b59a309cee.slice/crio-f47d90d516eca9a8057f5d5b24578c8ee9b5c880ba57bfe4167c45af570877ce WatchSource:0}: Error finding container f47d90d516eca9a8057f5d5b24578c8ee9b5c880ba57bfe4167c45af570877ce: Status 404 returned error can't find the container with id f47d90d516eca9a8057f5d5b24578c8ee9b5c880ba57bfe4167c45af570877ce Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.358100 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:04 crc kubenswrapper[4806]: E1204 03:57:04.358532 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:04.858510837 +0000 UTC m=+139.717023755 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.459476 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:04 crc kubenswrapper[4806]: E1204 03:57:04.459774 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:04.959759045 +0000 UTC m=+139.818271973 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.561241 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:04 crc kubenswrapper[4806]: E1204 03:57:04.561687 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:05.061671949 +0000 UTC m=+139.920184877 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.626885 4806 generic.go:334] "Generic (PLEG): container finished" podID="ba31bbad-4048-4651-92cd-d0cba1179541" containerID="7c4a29c3b94a01e72d4b805dc5f6e0e724ebb1c988b7fb34c5078fe55677f4d6" exitCode=0 Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.627379 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" event={"ID":"ba31bbad-4048-4651-92cd-d0cba1179541","Type":"ContainerDied","Data":"7c4a29c3b94a01e72d4b805dc5f6e0e724ebb1c988b7fb34c5078fe55677f4d6"} Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.659230 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-tclrk" event={"ID":"957c4786-76f4-4437-b56a-d42dee2c65b5","Type":"ContainerStarted","Data":"e2015ad621f83db87c029b1ece2d24dd603208120149e24d13c9cc7b6cca6a12"} Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.667919 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:04 crc kubenswrapper[4806]: E1204 03:57:04.669211 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:05.169190186 +0000 UTC m=+140.027703114 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.679511 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-mmxrh" event={"ID":"861e4822-a809-46c7-af83-06e8978ca8aa","Type":"ContainerStarted","Data":"d89be5b72c9f72dde0eb657a8a19591509e47d315ece1904f9c065fa745b74a1"} Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.735345 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh"] Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.744424 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" event={"ID":"b4f1c998-fb99-4827-a353-d8b59a309cee","Type":"ContainerStarted","Data":"f47d90d516eca9a8057f5d5b24578c8ee9b5c880ba57bfe4167c45af570877ce"} Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.773658 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:04 crc kubenswrapper[4806]: E1204 03:57:04.773951 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:05.273940186 +0000 UTC m=+140.132453104 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.784385 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-z2mdj"] Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.796361 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ln6xb" event={"ID":"3b036b1c-d79d-49d5-89ab-c669a5f8f47b","Type":"ContainerStarted","Data":"c13ff2012a4ce9fc29e03983908cac541c163614d34a52701f073fff338d03dc"} Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.872345 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" event={"ID":"920cdae1-9503-4181-92a2-d62969908d1d","Type":"ContainerStarted","Data":"942fbc152c0f3532953d696bac9a192f6e1631dcd430310b74d218e42a8d815d"} Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.875757 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:04 crc kubenswrapper[4806]: E1204 03:57:04.876179 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:05.37616121 +0000 UTC m=+140.234674138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.889767 4806 generic.go:334] "Generic (PLEG): container finished" podID="40553430-c0eb-45b4-880b-d0e1f85adb33" containerID="4c4eb74c696a7c2724133edbd670ef739550ac53c6c1817ad2a7ef9cf06411fc" exitCode=0 Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.889852 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" event={"ID":"40553430-c0eb-45b4-880b-d0e1f85adb33","Type":"ContainerDied","Data":"4c4eb74c696a7c2724133edbd670ef739550ac53c6c1817ad2a7ef9cf06411fc"} Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.899226 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" event={"ID":"606e54dc-5cc4-4e02-b796-dd37b613a8d0","Type":"ContainerStarted","Data":"6be8b06358583e17b0687a963e9bb13718431db5a968e305e3eb98c49e5ce081"} Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.899693 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.932213 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" podStartSLOduration=119.932197343 podStartE2EDuration="1m59.932197343s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:04.931638033 +0000 UTC m=+139.790150971" watchObservedRunningTime="2025-12-04 03:57:04.932197343 +0000 UTC m=+139.790710271" Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.961551 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" event={"ID":"7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6","Type":"ContainerStarted","Data":"b878efaaa45044d3f51998712113ace7824d42e528a3b780b45189729c9a806e"} Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.981799 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28"] Dec 04 03:57:04 crc kubenswrapper[4806]: I1204 03:57:04.985705 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:04 crc kubenswrapper[4806]: E1204 03:57:04.987949 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:05.487916564 +0000 UTC m=+140.346429492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.008260 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-j8m72" podStartSLOduration=120.008230697 podStartE2EDuration="2m0.008230697s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:04.994714999 +0000 UTC m=+139.853227937" watchObservedRunningTime="2025-12-04 03:57:05.008230697 +0000 UTC m=+139.866743635" Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.012179 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" event={"ID":"e9ff0349-32de-4d48-87ad-3e206ec56b2e","Type":"ContainerStarted","Data":"ef7a8c71e616271d3deac813bbcd56ba392864ce6bda7e825826dddfc733dd1d"} Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.016865 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn"] Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.016953 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f"] Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.030416 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p" event={"ID":"b5761a71-91cf-4781-9d7d-dc3f9b9096d0","Type":"ContainerStarted","Data":"378f2f8a13f090bb977a377991c1e965b03823e0b887b7fe36c12dc57e1e7ff0"} Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.041384 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-62l7r" event={"ID":"32ca5dd6-f353-4c80-ad4b-1dacf3057d1d","Type":"ContainerStarted","Data":"83812fe94801862a1d968e86dcc2fd585e5a829462ee7e1123ee079cb82fc479"} Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.042376 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-62l7r" Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.046499 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" event={"ID":"5229f915-0d18-41e7-abaf-1d2a783670eb","Type":"ContainerStarted","Data":"278b26a5fecc0768c84ae7d8f82ba4b2292fe9b40a7a8ac6d11e96e7bf69b6a7"} Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.048371 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" event={"ID":"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8","Type":"ContainerStarted","Data":"258c366591fe04778167b4b3595eb4316cd4f12b6070047abd9efcf1109f1470"} Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.049221 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.062494 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" event={"ID":"d10811c5-1f44-4668-a5f8-9690f2d81983","Type":"ContainerStarted","Data":"334ae323c491cec56e7deae81eb5c6c9cd421ce81cc8bdd398ce1ddbbd868c42"} Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.063627 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-z4k8p" podStartSLOduration=121.063610828 podStartE2EDuration="2m1.063610828s" podCreationTimestamp="2025-12-04 03:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:05.061437592 +0000 UTC m=+139.919950520" watchObservedRunningTime="2025-12-04 03:57:05.063610828 +0000 UTC m=+139.922123756" Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.074200 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dgkrj" event={"ID":"0647d456-31a3-4906-9d62-f43b4cfef5cb","Type":"ContainerStarted","Data":"e2854b663b19d14b9b307bc687ed5bd9c9f639400d2ca982586c2b9871f8df89"} Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.086498 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:05 crc kubenswrapper[4806]: E1204 03:57:05.087345 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:05.587317039 +0000 UTC m=+140.445829967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.087977 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" podStartSLOduration=120.087892709 podStartE2EDuration="2m0.087892709s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:05.084831883 +0000 UTC m=+139.943344821" watchObservedRunningTime="2025-12-04 03:57:05.087892709 +0000 UTC m=+139.946405637" Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.091417 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8nlkb" event={"ID":"5b4637a3-8d20-4427-8e71-9b5ab8e6ee8c","Type":"ContainerStarted","Data":"add3f65781a58214a8e74f71b745a3d376d8f04a33cf06ad3cc1008159ef4bd8"} Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.093981 4806 generic.go:334] "Generic (PLEG): container finished" podID="53e674b8-9044-45f1-8aa5-bb81ec63ae4a" containerID="9a325cc51d584023dd47858ac930777d7aa3bfc8406589139806687c6be40e2f" exitCode=0 Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.094038 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" event={"ID":"53e674b8-9044-45f1-8aa5-bb81ec63ae4a","Type":"ContainerDied","Data":"9a325cc51d584023dd47858ac930777d7aa3bfc8406589139806687c6be40e2f"} Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.094954 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8r5hx" event={"ID":"46000442-50bc-4480-9939-6a56032d3d2d","Type":"ContainerStarted","Data":"4d21ee2965fb58c5de45b5516107de6af906d0bc0b4aa16bc6e7da5b13550482"} Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.095762 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" event={"ID":"000c6858-a809-46e4-bdcb-b150477e5ce2","Type":"ContainerStarted","Data":"c02c42375f4dea63a29a4d64a761689fc676c47d9015e254a8afe60f84919977"} Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.096452 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" event={"ID":"79fbf847-050b-404d-8e9a-7f16f42089dc","Type":"ContainerStarted","Data":"718a22389c7b86ddb3451a7ab0bcea61d00167b7b0569ec5f5ee2ae1b4151894"} Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.099221 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db" event={"ID":"70d37d83-4fab-49db-8cb5-b98f50c02a0e","Type":"ContainerStarted","Data":"b12cd11d76785e074d3a2b81957c1659efc480f89209ebd950456702b8008865"} Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.099911 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv" event={"ID":"38a0c15f-4422-4119-b513-2e0ede6e8e8a","Type":"ContainerStarted","Data":"34f42ab2a1715885d6b03aafa389439d38470a9cd29bf360d7d2b0901d537b8c"} Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.101039 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-f47vx" event={"ID":"cdcade7f-c50d-4690-a339-db8dcc9a7ef0","Type":"ContainerStarted","Data":"4eebc1c3eb7cec23ca89ad4df55f0070504d1a548c3dc0a4bec689dc320115c5"} Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.107006 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-62l7r" podStartSLOduration=120.106992921 podStartE2EDuration="2m0.106992921s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:05.102697413 +0000 UTC m=+139.961210341" watchObservedRunningTime="2025-12-04 03:57:05.106992921 +0000 UTC m=+139.965505849" Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.127687 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-75rsr" podStartSLOduration=121.127651028 podStartE2EDuration="2m1.127651028s" podCreationTimestamp="2025-12-04 03:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:05.120314443 +0000 UTC m=+139.978827371" watchObservedRunningTime="2025-12-04 03:57:05.127651028 +0000 UTC m=+139.986163956" Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.188072 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:05 crc kubenswrapper[4806]: E1204 03:57:05.188374 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:05.688362902 +0000 UTC m=+140.546875830 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:05 crc kubenswrapper[4806]: W1204 03:57:05.220369 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc85a9a8a_89cf_4c60_9619_879f64d40195.slice/crio-2737717b164853fcb24480f7d3e2c4499d694f322303903ae902f30bcdb31153 WatchSource:0}: Error finding container 2737717b164853fcb24480f7d3e2c4499d694f322303903ae902f30bcdb31153: Status 404 returned error can't find the container with id 2737717b164853fcb24480f7d3e2c4499d694f322303903ae902f30bcdb31153 Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.220409 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.220465 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.277204 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-zp567"] Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.286404 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd"] Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.292476 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:05 crc kubenswrapper[4806]: E1204 03:57:05.292831 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:05.792815893 +0000 UTC m=+140.651328821 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.306335 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n"] Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.313619 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9"] Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.360280 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-62l7r" Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.393677 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:05 crc kubenswrapper[4806]: E1204 03:57:05.394022 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:05.894010511 +0000 UTC m=+140.752523439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.499661 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:05 crc kubenswrapper[4806]: E1204 03:57:05.500471 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:06.000452331 +0000 UTC m=+140.858965259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.500521 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:05 crc kubenswrapper[4806]: E1204 03:57:05.500861 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:06.000848134 +0000 UTC m=+140.859361062 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.517499 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bl2l4"] Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.578719 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4"] Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.605326 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:05 crc kubenswrapper[4806]: E1204 03:57:05.605640 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:06.105625106 +0000 UTC m=+140.964138034 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.617713 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cj594"] Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.626916 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mzgjp"] Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.706549 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:05 crc kubenswrapper[4806]: E1204 03:57:05.707426 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:06.207414965 +0000 UTC m=+141.065927893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.754001 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj"] Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.778705 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-tqxb8"] Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.789570 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8"] Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.802867 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vrgcp"] Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.808304 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:05 crc kubenswrapper[4806]: E1204 03:57:05.808732 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:06.308714386 +0000 UTC m=+141.167227314 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:05 crc kubenswrapper[4806]: I1204 03:57:05.909630 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:05 crc kubenswrapper[4806]: E1204 03:57:05.909963 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:06.409951865 +0000 UTC m=+141.268464793 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.012428 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:06 crc kubenswrapper[4806]: E1204 03:57:06.013059 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:06.513044329 +0000 UTC m=+141.371557257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.114051 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:06 crc kubenswrapper[4806]: E1204 03:57:06.114391 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:06.614379531 +0000 UTC m=+141.472892459 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.225096 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4" event={"ID":"33ec7e32-c944-4b4c-80c8-4f9a2ad49168","Type":"ContainerStarted","Data":"1ebf77ef99f5d67cc97845753a37561ab849d17e194c324c96e3a297a9da7edc"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.225581 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:06 crc kubenswrapper[4806]: E1204 03:57:06.226124 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:06.726108134 +0000 UTC m=+141.584621062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.251505 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n" event={"ID":"08021692-e26d-42ab-85a2-136b1bd2b4ed","Type":"ContainerStarted","Data":"c3487b3a516edb37115a3688420da5c8e44875a364561113dcb9b4e2d5937882"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.268733 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f" event={"ID":"1905e8ad-dfcf-4ce5-9565-28562f0832bc","Type":"ContainerStarted","Data":"827943adbc35bf0dbebd5e858cac682f0da338c3067e4876810720f4bd5d59af"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.270744 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" event={"ID":"9f43cb7f-513b-4958-88f6-e43a01c7adda","Type":"ContainerStarted","Data":"87ae6ad0f0e57270f850b504e111a83857b1ba6c392131294d7be7b5e495cfb1"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.329830 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" event={"ID":"000c6858-a809-46e4-bdcb-b150477e5ce2","Type":"ContainerStarted","Data":"2186f3a487247eca0a0153acc260f6eb90fd441fbe8e0427fc379a32eaa09295"} Dec 04 03:57:06 crc kubenswrapper[4806]: E1204 03:57:06.337632 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:06.83761163 +0000 UTC m=+141.696124558 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.337943 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.347364 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.355779 4806 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-7m42k container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" start-of-body= Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.355909 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" podUID="000c6858-a809-46e4-bdcb-b150477e5ce2" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.369492 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" event={"ID":"d10811c5-1f44-4668-a5f8-9690f2d81983","Type":"ContainerStarted","Data":"8ecd6570b633f38045df4076b5b2766de415f7a2366308566a7874006d052c63"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.391366 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-tqxb8" event={"ID":"e98c9985-31a9-4f14-818d-4519f0a3131b","Type":"ContainerStarted","Data":"de2bf2f6d0970cb96fe5e58dd4efc5b94067c76ec86b20c6f73db3cf5143be89"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.435470 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-mmxrh" event={"ID":"861e4822-a809-46c7-af83-06e8978ca8aa","Type":"ContainerStarted","Data":"5557ea872d735f3fbb1daff44733680c221d9ccd744e8c457ac73bbcbde54a35"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.444219 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:06 crc kubenswrapper[4806]: E1204 03:57:06.444411 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:06.944392341 +0000 UTC m=+141.802905269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.444586 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:06 crc kubenswrapper[4806]: E1204 03:57:06.445016 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:06.945000952 +0000 UTC m=+141.803513880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.446545 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bl2l4" event={"ID":"2f20b247-6932-459d-a19e-fd200ba45136","Type":"ContainerStarted","Data":"832fa2fd502e96e5fe37a0b3b23b6a0247d741564b347dab69f38144a102d1b2"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.457146 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-zp567" event={"ID":"a5cdccf3-8273-4553-abbf-fc0bedcf73fc","Type":"ContainerStarted","Data":"5dfe88d0019661c16d561b867af4164a1d4ed8ac7220def18295199e1a40bc6b"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.469737 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh" event={"ID":"c85a9a8a-89cf-4c60-9619-879f64d40195","Type":"ContainerStarted","Data":"c65c52510df6356613f4eb1362b5070f25346b4cb6a7f592830a98e861471bab"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.469791 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh" event={"ID":"c85a9a8a-89cf-4c60-9619-879f64d40195","Type":"ContainerStarted","Data":"2737717b164853fcb24480f7d3e2c4499d694f322303903ae902f30bcdb31153"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.545632 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:06 crc kubenswrapper[4806]: E1204 03:57:06.547527 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:07.047508046 +0000 UTC m=+141.906020994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.619185 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" event={"ID":"79fbf847-050b-404d-8e9a-7f16f42089dc","Type":"ContainerStarted","Data":"16ed8c1f70c984b51661a935d52d5cbb47a0f59e4200c11d1e68de2183d0cdd1"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.646076 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cj594" event={"ID":"5e02c1ac-06b2-4d89-98f8-545f682c4a4d","Type":"ContainerStarted","Data":"013afac0711434a37ed9e643a6b1dbd3096dd5f1b7c4056a454a3cb05b789e75"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.649385 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:06 crc kubenswrapper[4806]: E1204 03:57:06.649695 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:07.149681807 +0000 UTC m=+142.008194735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.651659 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db" event={"ID":"70d37d83-4fab-49db-8cb5-b98f50c02a0e","Type":"ContainerStarted","Data":"b096fd96aa20cccb08c87827e927098fb56f4c6d9e75f91ff97f209cfb334bd9"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.678778 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" event={"ID":"b5ee7463-4c14-4b15-974d-363744e4aaeb","Type":"ContainerStarted","Data":"ed390c4aebec1758ce93721c007ebade08f38d52340e3525469371f3694ecdb6"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.718070 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" event={"ID":"920cdae1-9503-4181-92a2-d62969908d1d","Type":"ContainerStarted","Data":"cbe00f220f597d8b956a443baaa9a2452969cbb62ef9dcfbf939d8a2f5fc4a3e"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.723808 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vrgcp" event={"ID":"4fbac3c5-6a92-4cc5-980d-5d66576df0e8","Type":"ContainerStarted","Data":"6d1925d0a06fa1e708f6addbe9339133920e144107dcd1685a18da7f0478aa84"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.729841 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-mmxrh" podStartSLOduration=121.729822755 podStartE2EDuration="2m1.729822755s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:06.728423837 +0000 UTC m=+141.586936785" watchObservedRunningTime="2025-12-04 03:57:06.729822755 +0000 UTC m=+141.588335683" Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.743407 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.750539 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.751308 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:06 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:06 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:06 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.751356 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.753728 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" event={"ID":"e9ff0349-32de-4d48-87ad-3e206ec56b2e","Type":"ContainerStarted","Data":"836f547c2c5a78b8da6f85029e9f5d12ada2fd2245079af2571755de0973ccb5"} Dec 04 03:57:06 crc kubenswrapper[4806]: E1204 03:57:06.752917 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:07.252898915 +0000 UTC m=+142.111411843 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.765764 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" event={"ID":"18bfd488-170c-457a-9abb-7b816f932bd0","Type":"ContainerStarted","Data":"94eea9b74d960f32362eaec2e833becf9fc13615be99f6bd639d6fb535ca9fd3"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.787446 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-db5db" podStartSLOduration=121.787428062 podStartE2EDuration="2m1.787428062s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:06.783792256 +0000 UTC m=+141.642305184" watchObservedRunningTime="2025-12-04 03:57:06.787428062 +0000 UTC m=+141.645940990" Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.791686 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv" event={"ID":"38a0c15f-4422-4119-b513-2e0ede6e8e8a","Type":"ContainerStarted","Data":"307dd3e70e75f9752e66277e742f85ebe8d4b3629e7f65832ef94b5fcce74c14"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.792675 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv" Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.806381 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dgkrj" event={"ID":"0647d456-31a3-4906-9d62-f43b4cfef5cb","Type":"ContainerStarted","Data":"0b44dcdb8220d2786d9fdf79ea56f93167c6522420206d16bd307fa4df529713"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.811832 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv" Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.840471 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-62ppz" podStartSLOduration=121.84044705 podStartE2EDuration="2m1.84044705s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:06.833340484 +0000 UTC m=+141.691853422" watchObservedRunningTime="2025-12-04 03:57:06.84044705 +0000 UTC m=+141.698959978" Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.845907 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8r5hx" event={"ID":"46000442-50bc-4480-9939-6a56032d3d2d","Type":"ContainerStarted","Data":"fef08f27994ed32965d9537710dc98e0a487f0d2ccf2224b47db739e5b057824"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.846354 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-8r5hx" Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.856160 4806 patch_prober.go:28] interesting pod/downloads-7954f5f757-8r5hx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.856229 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8r5hx" podUID="46000442-50bc-4480-9939-6a56032d3d2d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.857419 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:06 crc kubenswrapper[4806]: E1204 03:57:06.857753 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:07.3577411 +0000 UTC m=+142.216254028 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.905799 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-tclrk" event={"ID":"957c4786-76f4-4437-b56a-d42dee2c65b5","Type":"ContainerStarted","Data":"590ca8097f2a50236cac86e63395bc82bb848e21f496e04f4358b54b4dec020a"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.943521 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28" event={"ID":"f5fa998f-665b-421d-b589-e2531df3b2c2","Type":"ContainerStarted","Data":"b0c0f50c0acdd01b63c0a039180fddf0bf21f913f8044600e2bc2bb36fc4be24"} Dec 04 03:57:06 crc kubenswrapper[4806]: I1204 03:57:06.964514 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:06 crc kubenswrapper[4806]: E1204 03:57:06.966882 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:07.466863712 +0000 UTC m=+142.325376640 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.009016 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" event={"ID":"4e4767a8-ff51-43c7-94cc-625c9b8689a3","Type":"ContainerStarted","Data":"0d51f4cb252152e603656786b225cb1a33a5268ddddd3df2e38e9efff9648f5e"} Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.028958 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn" event={"ID":"7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e","Type":"ContainerStarted","Data":"93db3dfc63438d763e73ce9a7047e54e8eaba3c4b91dae3fbda4382499719088"} Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.039522 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-fd998" podStartSLOduration=122.03949324 podStartE2EDuration="2m2.03949324s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:06.945972878 +0000 UTC m=+141.804485806" watchObservedRunningTime="2025-12-04 03:57:07.03949324 +0000 UTC m=+141.898006178" Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.068392 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:07 crc kubenswrapper[4806]: E1204 03:57:07.070026 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:07.570015238 +0000 UTC m=+142.428528166 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.091549 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" podStartSLOduration=123.090991794 podStartE2EDuration="2m3.090991794s" podCreationTimestamp="2025-12-04 03:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:07.029728671 +0000 UTC m=+141.888241599" watchObservedRunningTime="2025-12-04 03:57:07.090991794 +0000 UTC m=+141.949504722" Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.091622 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8nlkb" event={"ID":"5b4637a3-8d20-4427-8e71-9b5ab8e6ee8c","Type":"ContainerStarted","Data":"74861c214cf36f281548001a7a91b38e17a134648694172a5a16abf166548a31"} Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.104893 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8" event={"ID":"fb71fca6-a96f-4438-84da-e1a6c7c58027","Type":"ContainerStarted","Data":"13907bc8c7cd0b628ba41992c45fcd80a6d7a86e303f03b833e1efbf798c8c3f"} Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.142709 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-z2mdj" event={"ID":"a3f377d2-10db-41db-9361-06abff2e0d29","Type":"ContainerStarted","Data":"283b3e846066c6ed47471e5a08b0db543f827e835d01a438071913c135915670"} Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.143026 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-z2mdj" event={"ID":"a3f377d2-10db-41db-9361-06abff2e0d29","Type":"ContainerStarted","Data":"ebf71e78512c42d1383b5fd5e7784665ac98983fced4976925e418cadb212c4e"} Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.143619 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28" podStartSLOduration=122.143606309 podStartE2EDuration="2m2.143606309s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:07.143581548 +0000 UTC m=+142.002094476" watchObservedRunningTime="2025-12-04 03:57:07.143606309 +0000 UTC m=+142.002119237" Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.147199 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nsqzh" podStartSLOduration=122.147189912 podStartE2EDuration="2m2.147189912s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:07.103531009 +0000 UTC m=+141.962043937" watchObservedRunningTime="2025-12-04 03:57:07.147189912 +0000 UTC m=+142.005702840" Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.173970 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:07 crc kubenswrapper[4806]: E1204 03:57:07.175415 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:07.67539469 +0000 UTC m=+142.533907638 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.221346 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dgkrj" podStartSLOduration=122.221328093 podStartE2EDuration="2m2.221328093s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:07.173557437 +0000 UTC m=+142.032070375" watchObservedRunningTime="2025-12-04 03:57:07.221328093 +0000 UTC m=+142.079841021" Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.279428 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:07 crc kubenswrapper[4806]: E1204 03:57:07.283294 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:07.7832752 +0000 UTC m=+142.641788208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.292639 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-6qzsg" podStartSLOduration=123.292615873 podStartE2EDuration="2m3.292615873s" podCreationTimestamp="2025-12-04 03:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:07.226281214 +0000 UTC m=+142.084794142" watchObservedRunningTime="2025-12-04 03:57:07.292615873 +0000 UTC m=+142.151128801" Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.353963 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-8r5hx" podStartSLOduration=122.35394354 podStartE2EDuration="2m2.35394354s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:07.294841921 +0000 UTC m=+142.153354849" watchObservedRunningTime="2025-12-04 03:57:07.35394354 +0000 UTC m=+142.212456468" Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.380612 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:07 crc kubenswrapper[4806]: E1204 03:57:07.381053 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:07.881036049 +0000 UTC m=+142.739548977 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.396035 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2c4tv" podStartSLOduration=122.396014909 podStartE2EDuration="2m2.396014909s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:07.355128201 +0000 UTC m=+142.213641129" watchObservedRunningTime="2025-12-04 03:57:07.396014909 +0000 UTC m=+142.254527827" Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.437869 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-tclrk" podStartSLOduration=8.437849539 podStartE2EDuration="8.437849539s" podCreationTimestamp="2025-12-04 03:56:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:07.397048924 +0000 UTC m=+142.255561872" watchObservedRunningTime="2025-12-04 03:57:07.437849539 +0000 UTC m=+142.296362477" Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.438124 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8nlkb" podStartSLOduration=123.438118847 podStartE2EDuration="2m3.438118847s" podCreationTimestamp="2025-12-04 03:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:07.436495621 +0000 UTC m=+142.295008539" watchObservedRunningTime="2025-12-04 03:57:07.438118847 +0000 UTC m=+142.296631775" Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.468948 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn" podStartSLOduration=122.468912234 podStartE2EDuration="2m2.468912234s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:07.468304224 +0000 UTC m=+142.326817152" watchObservedRunningTime="2025-12-04 03:57:07.468912234 +0000 UTC m=+142.327425162" Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.482126 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:07 crc kubenswrapper[4806]: E1204 03:57:07.482545 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:07.982531567 +0000 UTC m=+142.841044495 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.596601 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:07 crc kubenswrapper[4806]: E1204 03:57:07.597394 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:08.097379059 +0000 UTC m=+142.955891987 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.698755 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:07 crc kubenswrapper[4806]: E1204 03:57:07.699192 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:08.199176577 +0000 UTC m=+143.057689505 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.771634 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:07 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:07 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:07 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.771696 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.800346 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:07 crc kubenswrapper[4806]: E1204 03:57:07.800814 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:08.300793729 +0000 UTC m=+143.159306657 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:07 crc kubenswrapper[4806]: I1204 03:57:07.903838 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:07 crc kubenswrapper[4806]: E1204 03:57:07.904241 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:08.404226455 +0000 UTC m=+143.262739383 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.007423 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:08 crc kubenswrapper[4806]: E1204 03:57:08.007707 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:08.507679231 +0000 UTC m=+143.366192159 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.008151 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:08 crc kubenswrapper[4806]: E1204 03:57:08.008520 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:08.50850625 +0000 UTC m=+143.367019178 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.110463 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:08 crc kubenswrapper[4806]: E1204 03:57:08.111012 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:08.610992232 +0000 UTC m=+143.469505170 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.184332 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vrgcp" event={"ID":"4fbac3c5-6a92-4cc5-980d-5d66576df0e8","Type":"ContainerStarted","Data":"490b67f5b5da7729bd34fa45546a316c29306cbfeb6dff9e32ff63178e786ba8"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.194698 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" event={"ID":"9f43cb7f-513b-4958-88f6-e43a01c7adda","Type":"ContainerStarted","Data":"4a5c7c26475b8d7388e2671e841376a49e8bd8533978b7b2fbfe018a25e377bf"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.196320 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bl2l4" event={"ID":"2f20b247-6932-459d-a19e-fd200ba45136","Type":"ContainerStarted","Data":"0bf808a5553ba41e994aca4ed5ee3cd1a030d1e8429d469fd2870a7ea437555e"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.211474 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" event={"ID":"b4f1c998-fb99-4827-a353-d8b59a309cee","Type":"ContainerStarted","Data":"fa31bd1e6d4845d737a13097baa8f599989c9956742e0a3a2bc54315edef6652"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.213038 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:08 crc kubenswrapper[4806]: E1204 03:57:08.213377 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:08.713354371 +0000 UTC m=+143.571867299 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.247638 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4" event={"ID":"33ec7e32-c944-4b4c-80c8-4f9a2ad49168","Type":"ContainerStarted","Data":"6282eb6106ffdc5d829a8acc3c3d3be0237e74d9c86722864316528b53b688fa"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.283225 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-z2mdj" podStartSLOduration=123.283204092 podStartE2EDuration="2m3.283204092s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:07.549378944 +0000 UTC m=+142.407891882" watchObservedRunningTime="2025-12-04 03:57:08.283204092 +0000 UTC m=+143.141717030" Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.313917 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" event={"ID":"4e4767a8-ff51-43c7-94cc-625c9b8689a3","Type":"ContainerStarted","Data":"93979e497259d4fb735040c8bb60e0dbbf4d26aa5dad45718f9071972f3929ac"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.314220 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.315080 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" Dec 04 03:57:08 crc kubenswrapper[4806]: E1204 03:57:08.315411 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:08.815394237 +0000 UTC m=+143.673907225 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.319539 4806 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mzgjp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.319604 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" podUID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.324323 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-zp567" event={"ID":"a5cdccf3-8273-4553-abbf-fc0bedcf73fc","Type":"ContainerStarted","Data":"f4069666c58906629c78fa3433b0261a6b55e759ebe6229c2cb24b7f0faaad4d"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.326272 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" event={"ID":"18bfd488-170c-457a-9abb-7b816f932bd0","Type":"ContainerStarted","Data":"94599013cc59c7d1149f0cec354fa924c7a7ac58386939f58d3c97581232b99f"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.327127 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.328063 4806 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-6rqtd container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.328099 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" podUID="18bfd488-170c-457a-9abb-7b816f932bd0" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.328829 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" event={"ID":"b5ee7463-4c14-4b15-974d-363744e4aaeb","Type":"ContainerStarted","Data":"facce68f47a98a9ed0d23b8a60a637d71844e1b7265bfcdd4d9f81e6c4bcedde"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.329695 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.342880 4806 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-nn8tj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:5443/healthz\": dial tcp 10.217.0.21:5443: connect: connection refused" start-of-body= Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.342955 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" podUID="b5ee7463-4c14-4b15-974d-363744e4aaeb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.21:5443/healthz\": dial tcp 10.217.0.21:5443: connect: connection refused" Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.344482 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-djtqn" event={"ID":"7c4141fa-fe48-4c36-83f2-3b2dbf4eef5e","Type":"ContainerStarted","Data":"2ff8032fb77f84ad6abfb1b0c0a57047eab57f46ee0b5ee314deaf6a7a271f0e"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.350852 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f" event={"ID":"1905e8ad-dfcf-4ce5-9565-28562f0832bc","Type":"ContainerStarted","Data":"5968660b0c0437a5747bf604750b719de0dba129a0c5e05a6691008aa9820ff8"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.359556 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" event={"ID":"53e674b8-9044-45f1-8aa5-bb81ec63ae4a","Type":"ContainerStarted","Data":"8fee3ec67080a76ea22ca5b6d41c281b9c5395c7b903fbbfb83e9d1b9061426a"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.360269 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.389181 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" event={"ID":"ba31bbad-4048-4651-92cd-d0cba1179541","Type":"ContainerStarted","Data":"a55e0bb2fe3ab461326e25ae1ad25bc435dcb0dbf6c7bfec43bb3695bcc1b05e"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.401035 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" event={"ID":"ba31bbad-4048-4651-92cd-d0cba1179541","Type":"ContainerStarted","Data":"22817bcc637fd2dea0e72949c4a98cb1cffa2448cb0fb1332dd644a4fd78c4d2"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.401078 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8" event={"ID":"fb71fca6-a96f-4438-84da-e1a6c7c58027","Type":"ContainerStarted","Data":"02464d6251973483c87db0b46e5224d2f4eb30b7381b3ded70c5b5d7c86efa12"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.401094 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cvm28" event={"ID":"f5fa998f-665b-421d-b589-e2531df3b2c2","Type":"ContainerStarted","Data":"c7ab627c453d6a11ff65d56583807a6a02322f3810d85f951a4cd341a76cd7d5"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.401749 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" podStartSLOduration=123.40173411 podStartE2EDuration="2m3.40173411s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:08.293847081 +0000 UTC m=+143.152360009" watchObservedRunningTime="2025-12-04 03:57:08.40173411 +0000 UTC m=+143.260247038" Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.402854 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n" event={"ID":"08021692-e26d-42ab-85a2-136b1bd2b4ed","Type":"ContainerStarted","Data":"55f452bcddfb80fbdf8f6440895aa1aaf6ed5bf19b22cef627cc45aafa294f2a"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.417310 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:08 crc kubenswrapper[4806]: E1204 03:57:08.418907 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:08.918892285 +0000 UTC m=+143.777405213 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.438234 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-tqxb8" event={"ID":"e98c9985-31a9-4f14-818d-4519f0a3131b","Type":"ContainerStarted","Data":"59cf74a16631a052409363a896b22a01a3a60c3c8afb63639600ef632c9934a1"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.450202 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-f47vx" event={"ID":"cdcade7f-c50d-4690-a339-db8dcc9a7ef0","Type":"ContainerStarted","Data":"136b1039d1ba15091d76fe16e7bd2e9d4d13db6a4e757fe8d49471d49664a77d"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.450249 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-f47vx" event={"ID":"cdcade7f-c50d-4690-a339-db8dcc9a7ef0","Type":"ContainerStarted","Data":"67666d1dba79be8d66668efd6b8c7b397298141ed2f0a78cc376e75ddfc80cc4"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.450772 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-f47vx" Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.466114 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ln6xb" event={"ID":"3b036b1c-d79d-49d5-89ab-c669a5f8f47b","Type":"ContainerStarted","Data":"160218766ce2dda70ad7992db086714a553732ad409ec6ae09cbba065221e3cd"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.466167 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ln6xb" event={"ID":"3b036b1c-d79d-49d5-89ab-c669a5f8f47b","Type":"ContainerStarted","Data":"68e569875a40522c6c45c6dc3d81cc60e473f9efab02d6bb7da824858c3a0395"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.511268 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" event={"ID":"920cdae1-9503-4181-92a2-d62969908d1d","Type":"ContainerStarted","Data":"158f9f0bc419d351f50633eb76cc41c47f4e8735f4cbec589d9c3d86cd427fa5"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.514761 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" event={"ID":"40553430-c0eb-45b4-880b-d0e1f85adb33","Type":"ContainerStarted","Data":"701cf8e642e4873824fbe31e5f4163d692226512d68a20ad7d460a27ae5af07f"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.519975 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:08 crc kubenswrapper[4806]: E1204 03:57:08.521032 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:09.021017346 +0000 UTC m=+143.879530274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.537856 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cj594" event={"ID":"5e02c1ac-06b2-4d89-98f8-545f682c4a4d","Type":"ContainerStarted","Data":"21418d1de12f1e51f7aa5bace694d21cd876cd05cbc111e1feb22c8a9ffb1fe1"} Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.543108 4806 patch_prober.go:28] interesting pod/downloads-7954f5f757-8r5hx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.543187 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8r5hx" podUID="46000442-50bc-4480-9939-6a56032d3d2d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.547664 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" podStartSLOduration=123.547638508 podStartE2EDuration="2m3.547638508s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:08.401350997 +0000 UTC m=+143.259863935" watchObservedRunningTime="2025-12-04 03:57:08.547638508 +0000 UTC m=+143.406151436" Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.548194 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" podStartSLOduration=123.548187747 podStartE2EDuration="2m3.548187747s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:08.528136352 +0000 UTC m=+143.386649290" watchObservedRunningTime="2025-12-04 03:57:08.548187747 +0000 UTC m=+143.406700675" Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.623726 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:08 crc kubenswrapper[4806]: E1204 03:57:08.624852 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:09.124840495 +0000 UTC m=+143.983353413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.725014 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:08 crc kubenswrapper[4806]: E1204 03:57:08.726430 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:09.226402004 +0000 UTC m=+144.084914932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.748479 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:08 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:08 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:08 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.748726 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.763899 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" podStartSLOduration=124.763883064 podStartE2EDuration="2m4.763883064s" podCreationTimestamp="2025-12-04 03:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:08.74619049 +0000 UTC m=+143.604703418" watchObservedRunningTime="2025-12-04 03:57:08.763883064 +0000 UTC m=+143.622395982" Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.777031 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.827259 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:08 crc kubenswrapper[4806]: E1204 03:57:08.827642 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:09.327630463 +0000 UTC m=+144.186143391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.900907 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-7hgq8" podStartSLOduration=122.900874262 podStartE2EDuration="2m2.900874262s" podCreationTimestamp="2025-12-04 03:55:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:08.866310574 +0000 UTC m=+143.724823502" watchObservedRunningTime="2025-12-04 03:57:08.900874262 +0000 UTC m=+143.759387190" Dec 04 03:57:08 crc kubenswrapper[4806]: I1204 03:57:08.927837 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:08 crc kubenswrapper[4806]: E1204 03:57:08.928222 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:09.42820625 +0000 UTC m=+144.286719178 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.029897 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:09 crc kubenswrapper[4806]: E1204 03:57:09.030334 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:09.53031888 +0000 UTC m=+144.388831808 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.111204 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2fq8f" podStartSLOduration=124.111182763 podStartE2EDuration="2m4.111182763s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:09.010886016 +0000 UTC m=+143.869398944" watchObservedRunningTime="2025-12-04 03:57:09.111182763 +0000 UTC m=+143.969695691" Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.112679 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-zp567" podStartSLOduration=10.112672884 podStartE2EDuration="10.112672884s" podCreationTimestamp="2025-12-04 03:56:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:09.109443413 +0000 UTC m=+143.967956341" watchObservedRunningTime="2025-12-04 03:57:09.112672884 +0000 UTC m=+143.971185812" Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.131178 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:09 crc kubenswrapper[4806]: E1204 03:57:09.131471 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:09.631457716 +0000 UTC m=+144.489970644 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.232795 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:09 crc kubenswrapper[4806]: E1204 03:57:09.233166 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:09.73315345 +0000 UTC m=+144.591666378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.261513 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" podStartSLOduration=124.261490233 podStartE2EDuration="2m4.261490233s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:09.157651253 +0000 UTC m=+144.016164181" watchObservedRunningTime="2025-12-04 03:57:09.261490233 +0000 UTC m=+144.120003161" Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.333595 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:09 crc kubenswrapper[4806]: E1204 03:57:09.333837 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:09.833804969 +0000 UTC m=+144.692317907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.333898 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:09 crc kubenswrapper[4806]: E1204 03:57:09.334293 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:09.834271436 +0000 UTC m=+144.692784364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.334691 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9dhq6" podStartSLOduration=124.33467871 podStartE2EDuration="2m4.33467871s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:09.332982391 +0000 UTC m=+144.191495319" watchObservedRunningTime="2025-12-04 03:57:09.33467871 +0000 UTC m=+144.193191638" Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.337655 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" podStartSLOduration=124.337645923 podStartE2EDuration="2m4.337645923s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:09.264959883 +0000 UTC m=+144.123472811" watchObservedRunningTime="2025-12-04 03:57:09.337645923 +0000 UTC m=+144.196158851" Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.435136 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:09 crc kubenswrapper[4806]: E1204 03:57:09.435540 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:09.935517825 +0000 UTC m=+144.794030753 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.474084 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-f47vx" podStartSLOduration=10.474063591 podStartE2EDuration="10.474063591s" podCreationTimestamp="2025-12-04 03:56:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:09.468116515 +0000 UTC m=+144.326629443" watchObservedRunningTime="2025-12-04 03:57:09.474063591 +0000 UTC m=+144.332576519" Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.475744 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-ln6xb" podStartSLOduration=124.47573588 podStartE2EDuration="2m4.47573588s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:09.403282368 +0000 UTC m=+144.261795296" watchObservedRunningTime="2025-12-04 03:57:09.47573588 +0000 UTC m=+144.334248808" Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.536350 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:09 crc kubenswrapper[4806]: E1204 03:57:09.537005 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:10.036992213 +0000 UTC m=+144.895505141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.553335 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bl2l4" event={"ID":"2f20b247-6932-459d-a19e-fd200ba45136","Type":"ContainerStarted","Data":"f8481f404cd278230da397736d47da4beaa13bfb136b7e3927d12fe3619184a2"} Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.563592 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-tqxb8" podStartSLOduration=123.563574484 podStartE2EDuration="2m3.563574484s" podCreationTimestamp="2025-12-04 03:55:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:09.562396304 +0000 UTC m=+144.420909232" watchObservedRunningTime="2025-12-04 03:57:09.563574484 +0000 UTC m=+144.422087412" Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.608399 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4" event={"ID":"33ec7e32-c944-4b4c-80c8-4f9a2ad49168","Type":"ContainerStarted","Data":"2c7702e2ac1fe1cf5c2a7ca21c182f80e0b5c5b46a791694e0163aa7e3b9636d"} Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.609051 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4" Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.642810 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:09 crc kubenswrapper[4806]: E1204 03:57:09.643211 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:10.143196624 +0000 UTC m=+145.001709552 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.669330 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n" event={"ID":"08021692-e26d-42ab-85a2-136b1bd2b4ed","Type":"ContainerStarted","Data":"aa21b727d57f08080d7a2098c439ccdb70b73af37107164e81daf8795e701def"} Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.671592 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" podStartSLOduration=124.671574458 podStartE2EDuration="2m4.671574458s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:09.669436473 +0000 UTC m=+144.527949401" watchObservedRunningTime="2025-12-04 03:57:09.671574458 +0000 UTC m=+144.530087386" Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.679870 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vrgcp" event={"ID":"4fbac3c5-6a92-4cc5-980d-5d66576df0e8","Type":"ContainerStarted","Data":"d3700909e5beeac90afc810aae8a6abebe360c89a2a0e7c8473329d6ddad2d84"} Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.729425 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" event={"ID":"9f43cb7f-513b-4958-88f6-e43a01c7adda","Type":"ContainerStarted","Data":"11adc4d45687e970d47a73cc26a392062738bccc26d458e78f7c2283f4ba8ba6"} Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.731811 4806 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mzgjp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.731859 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" podUID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.745751 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:09 crc kubenswrapper[4806]: E1204 03:57:09.749663 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:10.249651245 +0000 UTC m=+145.108164163 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.754031 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:09 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:09 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:09 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.754061 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.769285 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6rqtd" Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.850437 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:09 crc kubenswrapper[4806]: E1204 03:57:09.851933 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:10.351903429 +0000 UTC m=+145.210416357 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.954652 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bl2l4" podStartSLOduration=124.95463619 podStartE2EDuration="2m4.95463619s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:09.849874469 +0000 UTC m=+144.708387387" watchObservedRunningTime="2025-12-04 03:57:09.95463619 +0000 UTC m=+144.813149118" Dec 04 03:57:09 crc kubenswrapper[4806]: I1204 03:57:09.956826 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:09 crc kubenswrapper[4806]: E1204 03:57:09.957154 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:10.457142087 +0000 UTC m=+145.315655015 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.057332 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:10 crc kubenswrapper[4806]: E1204 03:57:10.057536 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:10.557505546 +0000 UTC m=+145.416018474 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.057608 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:10 crc kubenswrapper[4806]: E1204 03:57:10.057905 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:10.55789282 +0000 UTC m=+145.416405748 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.101029 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wbb9n" podStartSLOduration=125.101006454 podStartE2EDuration="2m5.101006454s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:09.969890909 +0000 UTC m=+144.828403837" watchObservedRunningTime="2025-12-04 03:57:10.101006454 +0000 UTC m=+144.959519382" Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.158355 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:10 crc kubenswrapper[4806]: E1204 03:57:10.158538 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:10.658504957 +0000 UTC m=+145.517017885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.158665 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:10 crc kubenswrapper[4806]: E1204 03:57:10.159009 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:10.659001565 +0000 UTC m=+145.517514493 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.223771 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-vrgcp" podStartSLOduration=125.223752998 podStartE2EDuration="2m5.223752998s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:10.112991229 +0000 UTC m=+144.971504157" watchObservedRunningTime="2025-12-04 03:57:10.223752998 +0000 UTC m=+145.082265926" Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.224540 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4" podStartSLOduration=125.224534976 podStartE2EDuration="2m5.224534976s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:10.222604909 +0000 UTC m=+145.081117827" watchObservedRunningTime="2025-12-04 03:57:10.224534976 +0000 UTC m=+145.083047904" Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.232530 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jhptt" Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.259747 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:10 crc kubenswrapper[4806]: E1204 03:57:10.259975 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:10.759948184 +0000 UTC m=+145.618461112 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.260123 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:10 crc kubenswrapper[4806]: E1204 03:57:10.260469 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:10.760452421 +0000 UTC m=+145.618965349 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.362536 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:10 crc kubenswrapper[4806]: E1204 03:57:10.363297 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:10.863277586 +0000 UTC m=+145.721790524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.370848 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mgzl9" podStartSLOduration=125.370830727 podStartE2EDuration="2m5.370830727s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:10.303895467 +0000 UTC m=+145.162408405" watchObservedRunningTime="2025-12-04 03:57:10.370830727 +0000 UTC m=+145.229343655" Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.465583 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:10 crc kubenswrapper[4806]: E1204 03:57:10.465981 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:10.965969535 +0000 UTC m=+145.824482463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.566638 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:10 crc kubenswrapper[4806]: E1204 03:57:10.567259 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.067233895 +0000 UTC m=+145.925746823 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.668142 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:10 crc kubenswrapper[4806]: E1204 03:57:10.668448 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.168436174 +0000 UTC m=+146.026949102 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.729648 4806 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-nn8tj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.729727 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" podUID="b5ee7463-4c14-4b15-974d-363744e4aaeb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.21:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.737407 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cj594" event={"ID":"5e02c1ac-06b2-4d89-98f8-545f682c4a4d","Type":"ContainerStarted","Data":"0bf4f5c4d87dc5ddc39c2453787c8a0f69131d3a15085330f0120f765a409a2c"} Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.737463 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cj594" event={"ID":"5e02c1ac-06b2-4d89-98f8-545f682c4a4d","Type":"ContainerStarted","Data":"cda838e9ee02b5b7ff572f3a692838c0b45e8832d527d00468e2dc205f621d8e"} Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.741179 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.745576 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:10 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:10 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:10 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.745914 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.768971 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:10 crc kubenswrapper[4806]: E1204 03:57:10.769232 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.269184696 +0000 UTC m=+146.127697624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.769796 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:10 crc kubenswrapper[4806]: E1204 03:57:10.770168 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.270159629 +0000 UTC m=+146.128672627 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.842606 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q9fqz"] Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.843521 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.863998 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.867221 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q9fqz"] Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.870467 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.870601 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59a2e27f-ed83-42ba-9371-14546fbcc6af-catalog-content\") pod \"certified-operators-q9fqz\" (UID: \"59a2e27f-ed83-42ba-9371-14546fbcc6af\") " pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:57:10 crc kubenswrapper[4806]: E1204 03:57:10.870678 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.370650882 +0000 UTC m=+146.229163800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.870904 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.871061 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59a2e27f-ed83-42ba-9371-14546fbcc6af-utilities\") pod \"certified-operators-q9fqz\" (UID: \"59a2e27f-ed83-42ba-9371-14546fbcc6af\") " pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.871102 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqpw7\" (UniqueName: \"kubernetes.io/projected/59a2e27f-ed83-42ba-9371-14546fbcc6af-kube-api-access-pqpw7\") pod \"certified-operators-q9fqz\" (UID: \"59a2e27f-ed83-42ba-9371-14546fbcc6af\") " pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:57:10 crc kubenswrapper[4806]: E1204 03:57:10.871296 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.371283925 +0000 UTC m=+146.229796853 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.972133 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:10 crc kubenswrapper[4806]: E1204 03:57:10.972255 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.472238994 +0000 UTC m=+146.330751922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.972284 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59a2e27f-ed83-42ba-9371-14546fbcc6af-utilities\") pod \"certified-operators-q9fqz\" (UID: \"59a2e27f-ed83-42ba-9371-14546fbcc6af\") " pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.972312 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqpw7\" (UniqueName: \"kubernetes.io/projected/59a2e27f-ed83-42ba-9371-14546fbcc6af-kube-api-access-pqpw7\") pod \"certified-operators-q9fqz\" (UID: \"59a2e27f-ed83-42ba-9371-14546fbcc6af\") " pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.972349 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59a2e27f-ed83-42ba-9371-14546fbcc6af-catalog-content\") pod \"certified-operators-q9fqz\" (UID: \"59a2e27f-ed83-42ba-9371-14546fbcc6af\") " pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.972406 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:10 crc kubenswrapper[4806]: E1204 03:57:10.972667 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.472648898 +0000 UTC m=+146.331161816 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.972780 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59a2e27f-ed83-42ba-9371-14546fbcc6af-utilities\") pod \"certified-operators-q9fqz\" (UID: \"59a2e27f-ed83-42ba-9371-14546fbcc6af\") " pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:57:10 crc kubenswrapper[4806]: I1204 03:57:10.973046 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59a2e27f-ed83-42ba-9371-14546fbcc6af-catalog-content\") pod \"certified-operators-q9fqz\" (UID: \"59a2e27f-ed83-42ba-9371-14546fbcc6af\") " pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.013266 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rlrnj"] Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.014467 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.038654 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.058208 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqpw7\" (UniqueName: \"kubernetes.io/projected/59a2e27f-ed83-42ba-9371-14546fbcc6af-kube-api-access-pqpw7\") pod \"certified-operators-q9fqz\" (UID: \"59a2e27f-ed83-42ba-9371-14546fbcc6af\") " pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.073728 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:11 crc kubenswrapper[4806]: E1204 03:57:11.073932 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.573890038 +0000 UTC m=+146.432402966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.074121 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.074182 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21c59133-1783-4225-a70d-1395d9642980-catalog-content\") pod \"community-operators-rlrnj\" (UID: \"21c59133-1783-4225-a70d-1395d9642980\") " pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.074225 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffvww\" (UniqueName: \"kubernetes.io/projected/21c59133-1783-4225-a70d-1395d9642980-kube-api-access-ffvww\") pod \"community-operators-rlrnj\" (UID: \"21c59133-1783-4225-a70d-1395d9642980\") " pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.074388 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21c59133-1783-4225-a70d-1395d9642980-utilities\") pod \"community-operators-rlrnj\" (UID: \"21c59133-1783-4225-a70d-1395d9642980\") " pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:57:11 crc kubenswrapper[4806]: E1204 03:57:11.074451 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.574444257 +0000 UTC m=+146.432957185 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.091821 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rlrnj"] Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.156021 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.177523 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.177754 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffvww\" (UniqueName: \"kubernetes.io/projected/21c59133-1783-4225-a70d-1395d9642980-kube-api-access-ffvww\") pod \"community-operators-rlrnj\" (UID: \"21c59133-1783-4225-a70d-1395d9642980\") " pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.177841 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21c59133-1783-4225-a70d-1395d9642980-utilities\") pod \"community-operators-rlrnj\" (UID: \"21c59133-1783-4225-a70d-1395d9642980\") " pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.178011 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21c59133-1783-4225-a70d-1395d9642980-catalog-content\") pod \"community-operators-rlrnj\" (UID: \"21c59133-1783-4225-a70d-1395d9642980\") " pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.179071 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21c59133-1783-4225-a70d-1395d9642980-catalog-content\") pod \"community-operators-rlrnj\" (UID: \"21c59133-1783-4225-a70d-1395d9642980\") " pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:57:11 crc kubenswrapper[4806]: E1204 03:57:11.179268 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.67925076 +0000 UTC m=+146.537763688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.179519 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21c59133-1783-4225-a70d-1395d9642980-utilities\") pod \"community-operators-rlrnj\" (UID: \"21c59133-1783-4225-a70d-1395d9642980\") " pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.220358 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-brh2b"] Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.221593 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.279155 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d1500e-9acd-47da-b1cb-13a7c7123f2a-catalog-content\") pod \"certified-operators-brh2b\" (UID: \"07d1500e-9acd-47da-b1cb-13a7c7123f2a\") " pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.279209 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.279242 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnddt\" (UniqueName: \"kubernetes.io/projected/07d1500e-9acd-47da-b1cb-13a7c7123f2a-kube-api-access-dnddt\") pod \"certified-operators-brh2b\" (UID: \"07d1500e-9acd-47da-b1cb-13a7c7123f2a\") " pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.279270 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.279321 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d1500e-9acd-47da-b1cb-13a7c7123f2a-utilities\") pod \"certified-operators-brh2b\" (UID: \"07d1500e-9acd-47da-b1cb-13a7c7123f2a\") " pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.279341 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:57:11 crc kubenswrapper[4806]: E1204 03:57:11.281294 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.781279847 +0000 UTC m=+146.639792775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.284709 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.296615 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.380107 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.380425 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnddt\" (UniqueName: \"kubernetes.io/projected/07d1500e-9acd-47da-b1cb-13a7c7123f2a-kube-api-access-dnddt\") pod \"certified-operators-brh2b\" (UID: \"07d1500e-9acd-47da-b1cb-13a7c7123f2a\") " pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.380480 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.380541 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d1500e-9acd-47da-b1cb-13a7c7123f2a-utilities\") pod \"certified-operators-brh2b\" (UID: \"07d1500e-9acd-47da-b1cb-13a7c7123f2a\") " pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.380573 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.380609 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d1500e-9acd-47da-b1cb-13a7c7123f2a-catalog-content\") pod \"certified-operators-brh2b\" (UID: \"07d1500e-9acd-47da-b1cb-13a7c7123f2a\") " pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.381115 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d1500e-9acd-47da-b1cb-13a7c7123f2a-catalog-content\") pod \"certified-operators-brh2b\" (UID: \"07d1500e-9acd-47da-b1cb-13a7c7123f2a\") " pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:57:11 crc kubenswrapper[4806]: E1204 03:57:11.381197 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.8811798 +0000 UTC m=+146.739692728 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.389632 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d1500e-9acd-47da-b1cb-13a7c7123f2a-utilities\") pod \"certified-operators-brh2b\" (UID: \"07d1500e-9acd-47da-b1cb-13a7c7123f2a\") " pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.396368 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.399428 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-brh2b"] Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.410949 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-74mh7"] Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.411855 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.464702 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-74mh7"] Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.482192 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:11 crc kubenswrapper[4806]: E1204 03:57:11.482573 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:11.982561074 +0000 UTC m=+146.841074002 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.483395 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnddt\" (UniqueName: \"kubernetes.io/projected/07d1500e-9acd-47da-b1cb-13a7c7123f2a-kube-api-access-dnddt\") pod \"certified-operators-brh2b\" (UID: \"07d1500e-9acd-47da-b1cb-13a7c7123f2a\") " pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.550652 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.555198 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.568793 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.583336 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.583760 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85470b07-1f66-4a8a-b30c-85c2728ab076-utilities\") pod \"community-operators-74mh7\" (UID: \"85470b07-1f66-4a8a-b30c-85c2728ab076\") " pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.583847 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85470b07-1f66-4a8a-b30c-85c2728ab076-catalog-content\") pod \"community-operators-74mh7\" (UID: \"85470b07-1f66-4a8a-b30c-85c2728ab076\") " pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.583954 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66bms\" (UniqueName: \"kubernetes.io/projected/85470b07-1f66-4a8a-b30c-85c2728ab076-kube-api-access-66bms\") pod \"community-operators-74mh7\" (UID: \"85470b07-1f66-4a8a-b30c-85c2728ab076\") " pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:57:11 crc kubenswrapper[4806]: E1204 03:57:11.584380 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:12.084360782 +0000 UTC m=+146.942873710 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.603694 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.604907 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.679111 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.683994 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffvww\" (UniqueName: \"kubernetes.io/projected/21c59133-1783-4225-a70d-1395d9642980-kube-api-access-ffvww\") pod \"community-operators-rlrnj\" (UID: \"21c59133-1783-4225-a70d-1395d9642980\") " pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.684727 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.684787 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66bms\" (UniqueName: \"kubernetes.io/projected/85470b07-1f66-4a8a-b30c-85c2728ab076-kube-api-access-66bms\") pod \"community-operators-74mh7\" (UID: \"85470b07-1f66-4a8a-b30c-85c2728ab076\") " pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.684842 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85470b07-1f66-4a8a-b30c-85c2728ab076-utilities\") pod \"community-operators-74mh7\" (UID: \"85470b07-1f66-4a8a-b30c-85c2728ab076\") " pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.684903 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85470b07-1f66-4a8a-b30c-85c2728ab076-catalog-content\") pod \"community-operators-74mh7\" (UID: \"85470b07-1f66-4a8a-b30c-85c2728ab076\") " pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.685385 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85470b07-1f66-4a8a-b30c-85c2728ab076-utilities\") pod \"community-operators-74mh7\" (UID: \"85470b07-1f66-4a8a-b30c-85c2728ab076\") " pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:57:11 crc kubenswrapper[4806]: E1204 03:57:11.685522 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:12.185507839 +0000 UTC m=+147.044020767 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.686213 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85470b07-1f66-4a8a-b30c-85c2728ab076-catalog-content\") pod \"community-operators-74mh7\" (UID: \"85470b07-1f66-4a8a-b30c-85c2728ab076\") " pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.719687 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66bms\" (UniqueName: \"kubernetes.io/projected/85470b07-1f66-4a8a-b30c-85c2728ab076-kube-api-access-66bms\") pod \"community-operators-74mh7\" (UID: \"85470b07-1f66-4a8a-b30c-85c2728ab076\") " pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.739616 4806 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-nn8tj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.739695 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" podUID="b5ee7463-4c14-4b15-974d-363744e4aaeb" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.21:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.747709 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:11 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:11 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:11 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.747774 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.786502 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:11 crc kubenswrapper[4806]: E1204 03:57:11.786670 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:12.286643405 +0000 UTC m=+147.145156333 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.787649 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:11 crc kubenswrapper[4806]: E1204 03:57:11.790492 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:12.290476408 +0000 UTC m=+147.148989526 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.804015 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.823232 4806 patch_prober.go:28] interesting pod/apiserver-76f77b778f-7lbq7 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 04 03:57:11 crc kubenswrapper[4806]: [+]log ok Dec 04 03:57:11 crc kubenswrapper[4806]: [+]etcd ok Dec 04 03:57:11 crc kubenswrapper[4806]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 04 03:57:11 crc kubenswrapper[4806]: [+]poststarthook/generic-apiserver-start-informers ok Dec 04 03:57:11 crc kubenswrapper[4806]: [+]poststarthook/max-in-flight-filter ok Dec 04 03:57:11 crc kubenswrapper[4806]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 04 03:57:11 crc kubenswrapper[4806]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 04 03:57:11 crc kubenswrapper[4806]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 04 03:57:11 crc kubenswrapper[4806]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 04 03:57:11 crc kubenswrapper[4806]: [+]poststarthook/project.openshift.io-projectcache ok Dec 04 03:57:11 crc kubenswrapper[4806]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 04 03:57:11 crc kubenswrapper[4806]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Dec 04 03:57:11 crc kubenswrapper[4806]: [-]poststarthook/openshift.io-restmapperupdater failed: reason withheld Dec 04 03:57:11 crc kubenswrapper[4806]: [-]poststarthook/quota.openshift.io-clusterquotamapping failed: reason withheld Dec 04 03:57:11 crc kubenswrapper[4806]: livez check failed Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.823299 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" podUID="ba31bbad-4048-4651-92cd-d0cba1179541" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.829144 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.829186 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.847156 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.854405 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.888615 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:11 crc kubenswrapper[4806]: E1204 03:57:11.889117 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:12.389078626 +0000 UTC m=+147.247591554 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.922568 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q9fqz"] Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.931248 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:57:11 crc kubenswrapper[4806]: W1204 03:57:11.976760 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59a2e27f_ed83_42ba_9371_14546fbcc6af.slice/crio-f2807b5b634ebb72f89240c81f551a1a4cde072bf9a8bc29b22fab7acd4852ca WatchSource:0}: Error finding container f2807b5b634ebb72f89240c81f551a1a4cde072bf9a8bc29b22fab7acd4852ca: Status 404 returned error can't find the container with id f2807b5b634ebb72f89240c81f551a1a4cde072bf9a8bc29b22fab7acd4852ca Dec 04 03:57:11 crc kubenswrapper[4806]: I1204 03:57:11.993661 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:11 crc kubenswrapper[4806]: E1204 03:57:11.993960 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:12.493948271 +0000 UTC m=+147.352461199 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.095489 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:12 crc kubenswrapper[4806]: E1204 03:57:12.095875 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:12.595853453 +0000 UTC m=+147.454366381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.197083 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:12 crc kubenswrapper[4806]: E1204 03:57:12.197453 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:12.697436944 +0000 UTC m=+147.555949872 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.298541 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:12 crc kubenswrapper[4806]: E1204 03:57:12.298764 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:12.798736936 +0000 UTC m=+147.657249864 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.299211 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:12 crc kubenswrapper[4806]: E1204 03:57:12.299604 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:12.799592916 +0000 UTC m=+147.658105844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.397486 4806 patch_prober.go:28] interesting pod/downloads-7954f5f757-8r5hx container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.397547 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8r5hx" podUID="46000442-50bc-4480-9939-6a56032d3d2d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.397716 4806 patch_prober.go:28] interesting pod/downloads-7954f5f757-8r5hx container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.397761 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8r5hx" podUID="46000442-50bc-4480-9939-6a56032d3d2d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.401814 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:12 crc kubenswrapper[4806]: E1204 03:57:12.402116 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:12.902068387 +0000 UTC m=+147.760581315 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.403044 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:12 crc kubenswrapper[4806]: E1204 03:57:12.403441 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:12.903430195 +0000 UTC m=+147.761943123 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.504513 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:12 crc kubenswrapper[4806]: E1204 03:57:12.505115 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:13.005075809 +0000 UTC m=+147.863588727 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.609618 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:12 crc kubenswrapper[4806]: E1204 03:57:12.609922 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:13.109907442 +0000 UTC m=+147.968420360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.678627 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-74mh7"] Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.711662 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:12 crc kubenswrapper[4806]: E1204 03:57:12.711965 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:13.21194973 +0000 UTC m=+148.070462658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.745162 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.763552 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:12 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:12 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:12 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.763604 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.795031 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.796012 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.800069 4806 patch_prober.go:28] interesting pod/console-f9d7485db-z2mdj container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.800118 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-z2mdj" podUID="a3f377d2-10db-41db-9361-06abff2e0d29" containerName="console" probeResult="failure" output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.812597 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.813116 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"afd2163e12ddf66c055f2def2b4fd30818ebe010bcd971cfd177b6fe8fdca7b3"} Dec 04 03:57:12 crc kubenswrapper[4806]: E1204 03:57:12.813388 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:13.313377195 +0000 UTC m=+148.171890123 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.837195 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-74mh7" event={"ID":"85470b07-1f66-4a8a-b30c-85c2728ab076","Type":"ContainerStarted","Data":"5d6a3c5e060d7b05a8b056252ec00c3383e33146bb35cbc6b5ed05263fbc1cad"} Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.849910 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nn8tj" Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.865328 4806 generic.go:334] "Generic (PLEG): container finished" podID="b4f1c998-fb99-4827-a353-d8b59a309cee" containerID="fa31bd1e6d4845d737a13097baa8f599989c9956742e0a3a2bc54315edef6652" exitCode=0 Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.865398 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" event={"ID":"b4f1c998-fb99-4827-a353-d8b59a309cee","Type":"ContainerDied","Data":"fa31bd1e6d4845d737a13097baa8f599989c9956742e0a3a2bc54315edef6652"} Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.885298 4806 generic.go:334] "Generic (PLEG): container finished" podID="59a2e27f-ed83-42ba-9371-14546fbcc6af" containerID="a41fbc6a04d874c280f0a8062014a00be43ae847613fb859d990c740824fe46b" exitCode=0 Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.886044 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q9fqz" event={"ID":"59a2e27f-ed83-42ba-9371-14546fbcc6af","Type":"ContainerDied","Data":"a41fbc6a04d874c280f0a8062014a00be43ae847613fb859d990c740824fe46b"} Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.886074 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q9fqz" event={"ID":"59a2e27f-ed83-42ba-9371-14546fbcc6af","Type":"ContainerStarted","Data":"f2807b5b634ebb72f89240c81f551a1a4cde072bf9a8bc29b22fab7acd4852ca"} Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.895643 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.920051 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:12 crc kubenswrapper[4806]: E1204 03:57:12.921727 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:13.421707101 +0000 UTC m=+148.280220029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.930830 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cj594" event={"ID":"5e02c1ac-06b2-4d89-98f8-545f682c4a4d","Type":"ContainerStarted","Data":"f484f4bf899ff1875b6802da99bbc0ff383d0bc53f0011816693a0483e7f1fa8"} Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.941416 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rlrnj"] Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.963915 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g89mj" Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.974519 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ln2r5"] Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.974710 4806 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.975845 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:57:12 crc kubenswrapper[4806]: I1204 03:57:12.979403 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 03:57:12 crc kubenswrapper[4806]: W1204 03:57:12.982496 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21c59133_1783_4225_a70d_1395d9642980.slice/crio-6d8e220fedeb7bcedbe609594e0264e951bb2760975027156d0b72c2118d7466 WatchSource:0}: Error finding container 6d8e220fedeb7bcedbe609594e0264e951bb2760975027156d0b72c2118d7466: Status 404 returned error can't find the container with id 6d8e220fedeb7bcedbe609594e0264e951bb2760975027156d0b72c2118d7466 Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.021577 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-catalog-content\") pod \"redhat-marketplace-ln2r5\" (UID: \"ef7e1243-9a57-4bbc-a28a-43cb93cf2add\") " pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.021626 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf6x7\" (UniqueName: \"kubernetes.io/projected/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-kube-api-access-wf6x7\") pod \"redhat-marketplace-ln2r5\" (UID: \"ef7e1243-9a57-4bbc-a28a-43cb93cf2add\") " pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.021655 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.021693 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-utilities\") pod \"redhat-marketplace-ln2r5\" (UID: \"ef7e1243-9a57-4bbc-a28a-43cb93cf2add\") " pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:57:13 crc kubenswrapper[4806]: E1204 03:57:13.022314 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:13.522298278 +0000 UTC m=+148.380811206 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.031989 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-brh2b"] Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.070992 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ln2r5"] Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.123003 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.123435 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-utilities\") pod \"redhat-marketplace-ln2r5\" (UID: \"ef7e1243-9a57-4bbc-a28a-43cb93cf2add\") " pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.123516 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-catalog-content\") pod \"redhat-marketplace-ln2r5\" (UID: \"ef7e1243-9a57-4bbc-a28a-43cb93cf2add\") " pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.123542 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf6x7\" (UniqueName: \"kubernetes.io/projected/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-kube-api-access-wf6x7\") pod \"redhat-marketplace-ln2r5\" (UID: \"ef7e1243-9a57-4bbc-a28a-43cb93cf2add\") " pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:57:13 crc kubenswrapper[4806]: E1204 03:57:13.124731 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:13.624700597 +0000 UTC m=+148.483213525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.125146 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-utilities\") pod \"redhat-marketplace-ln2r5\" (UID: \"ef7e1243-9a57-4bbc-a28a-43cb93cf2add\") " pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.125358 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-catalog-content\") pod \"redhat-marketplace-ln2r5\" (UID: \"ef7e1243-9a57-4bbc-a28a-43cb93cf2add\") " pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.194188 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf6x7\" (UniqueName: \"kubernetes.io/projected/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-kube-api-access-wf6x7\") pod \"redhat-marketplace-ln2r5\" (UID: \"ef7e1243-9a57-4bbc-a28a-43cb93cf2add\") " pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.196045 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-cj594" podStartSLOduration=14.195980978 podStartE2EDuration="14.195980978s" podCreationTimestamp="2025-12-04 03:56:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:13.194640701 +0000 UTC m=+148.053153639" watchObservedRunningTime="2025-12-04 03:57:13.195980978 +0000 UTC m=+148.054493906" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.226692 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:13 crc kubenswrapper[4806]: E1204 03:57:13.227105 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:13.727092227 +0000 UTC m=+148.585605155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.308285 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.327802 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:13 crc kubenswrapper[4806]: E1204 03:57:13.328203 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:13.828184751 +0000 UTC m=+148.686697669 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.381511 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zjp4q"] Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.383432 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.398176 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjp4q"] Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.428686 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.428782 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1951d0-8637-4110-ba57-ac3ee5f6c378-catalog-content\") pod \"redhat-marketplace-zjp4q\" (UID: \"8d1951d0-8637-4110-ba57-ac3ee5f6c378\") " pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.428821 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4574\" (UniqueName: \"kubernetes.io/projected/8d1951d0-8637-4110-ba57-ac3ee5f6c378-kube-api-access-h4574\") pod \"redhat-marketplace-zjp4q\" (UID: \"8d1951d0-8637-4110-ba57-ac3ee5f6c378\") " pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.428838 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1951d0-8637-4110-ba57-ac3ee5f6c378-utilities\") pod \"redhat-marketplace-zjp4q\" (UID: \"8d1951d0-8637-4110-ba57-ac3ee5f6c378\") " pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:57:13 crc kubenswrapper[4806]: E1204 03:57:13.429176 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:13.929164702 +0000 UTC m=+148.787677630 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.529616 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:13 crc kubenswrapper[4806]: E1204 03:57:13.529733 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:14.029717877 +0000 UTC m=+148.888230805 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.530076 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1951d0-8637-4110-ba57-ac3ee5f6c378-catalog-content\") pod \"redhat-marketplace-zjp4q\" (UID: \"8d1951d0-8637-4110-ba57-ac3ee5f6c378\") " pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.530129 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4574\" (UniqueName: \"kubernetes.io/projected/8d1951d0-8637-4110-ba57-ac3ee5f6c378-kube-api-access-h4574\") pod \"redhat-marketplace-zjp4q\" (UID: \"8d1951d0-8637-4110-ba57-ac3ee5f6c378\") " pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.530149 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1951d0-8637-4110-ba57-ac3ee5f6c378-utilities\") pod \"redhat-marketplace-zjp4q\" (UID: \"8d1951d0-8637-4110-ba57-ac3ee5f6c378\") " pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.530191 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:13 crc kubenswrapper[4806]: E1204 03:57:13.530531 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:14.030523304 +0000 UTC m=+148.889036232 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.531231 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1951d0-8637-4110-ba57-ac3ee5f6c378-catalog-content\") pod \"redhat-marketplace-zjp4q\" (UID: \"8d1951d0-8637-4110-ba57-ac3ee5f6c378\") " pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.532439 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1951d0-8637-4110-ba57-ac3ee5f6c378-utilities\") pod \"redhat-marketplace-zjp4q\" (UID: \"8d1951d0-8637-4110-ba57-ac3ee5f6c378\") " pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.580477 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4574\" (UniqueName: \"kubernetes.io/projected/8d1951d0-8637-4110-ba57-ac3ee5f6c378-kube-api-access-h4574\") pod \"redhat-marketplace-zjp4q\" (UID: \"8d1951d0-8637-4110-ba57-ac3ee5f6c378\") " pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.634512 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:13 crc kubenswrapper[4806]: E1204 03:57:13.634999 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:14.134961135 +0000 UTC m=+148.993474073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.646411 4806 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-04T03:57:12.974736669Z","Handler":null,"Name":""} Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.736290 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.736911 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:13 crc kubenswrapper[4806]: E1204 03:57:13.737301 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:14.237288081 +0000 UTC m=+149.095801009 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.746030 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:13 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:13 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:13 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.746093 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.839154 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:13 crc kubenswrapper[4806]: E1204 03:57:13.839392 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 03:57:14.33935561 +0000 UTC m=+149.197868538 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.839460 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:13 crc kubenswrapper[4806]: E1204 03:57:13.840128 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 03:57:14.340118707 +0000 UTC m=+149.198631635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c2d5j" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.851365 4806 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.851407 4806 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.923242 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ln2r5"] Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.941068 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.953592 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hwdhx"] Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.955283 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.963130 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 03:57:13 crc kubenswrapper[4806]: I1204 03:57:13.981902 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hwdhx"] Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.002117 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 03:57:14 crc kubenswrapper[4806]: W1204 03:57:14.015903 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef7e1243_9a57_4bbc_a28a_43cb93cf2add.slice/crio-99116f9e4ba44a8de33bc4f00aaf26a6059a9a5d23db206cc189bc6b343fec9d WatchSource:0}: Error finding container 99116f9e4ba44a8de33bc4f00aaf26a6059a9a5d23db206cc189bc6b343fec9d: Status 404 returned error can't find the container with id 99116f9e4ba44a8de33bc4f00aaf26a6059a9a5d23db206cc189bc6b343fec9d Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.037667 4806 generic.go:334] "Generic (PLEG): container finished" podID="07d1500e-9acd-47da-b1cb-13a7c7123f2a" containerID="8de6a4d31028447c381ab0ba5d16e8b653b96a190f095f9886548410ba66738f" exitCode=0 Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.037724 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brh2b" event={"ID":"07d1500e-9acd-47da-b1cb-13a7c7123f2a","Type":"ContainerDied","Data":"8de6a4d31028447c381ab0ba5d16e8b653b96a190f095f9886548410ba66738f"} Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.037751 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brh2b" event={"ID":"07d1500e-9acd-47da-b1cb-13a7c7123f2a","Type":"ContainerStarted","Data":"e9a0398b52fa1880967552bfc2050383ed3cfed471193fa2462d9a1372f4ba9c"} Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.043604 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaca0644-a305-41b3-84f1-4179e3412d37-catalog-content\") pod \"redhat-operators-hwdhx\" (UID: \"eaca0644-a305-41b3-84f1-4179e3412d37\") " pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.043730 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaca0644-a305-41b3-84f1-4179e3412d37-utilities\") pod \"redhat-operators-hwdhx\" (UID: \"eaca0644-a305-41b3-84f1-4179e3412d37\") " pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.043828 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.044997 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94gnq\" (UniqueName: \"kubernetes.io/projected/eaca0644-a305-41b3-84f1-4179e3412d37-kube-api-access-94gnq\") pod \"redhat-operators-hwdhx\" (UID: \"eaca0644-a305-41b3-84f1-4179e3412d37\") " pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.044659 4806 generic.go:334] "Generic (PLEG): container finished" podID="21c59133-1783-4225-a70d-1395d9642980" containerID="190950118eae2c27a4c54356561942af9725ab4f3cb228201e31e08d44905dd7" exitCode=0 Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.048773 4806 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.048803 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.044724 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlrnj" event={"ID":"21c59133-1783-4225-a70d-1395d9642980","Type":"ContainerDied","Data":"190950118eae2c27a4c54356561942af9725ab4f3cb228201e31e08d44905dd7"} Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.061163 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlrnj" event={"ID":"21c59133-1783-4225-a70d-1395d9642980","Type":"ContainerStarted","Data":"6d8e220fedeb7bcedbe609594e0264e951bb2760975027156d0b72c2118d7466"} Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.085779 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"e67fe60fddff448b3d2ee64c9c93b911c9b5f43360eb24d598c330d80f8db020"} Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.085839 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a56b6e05e1a5417dcf76d78945778b3b882dba8ab9dc2e29653acace4537653b"} Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.097988 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"20ae19af0cae530dde75dae7e016e5a9af292897167f61461fa5d15842f44269"} Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.110649 4806 generic.go:334] "Generic (PLEG): container finished" podID="85470b07-1f66-4a8a-b30c-85c2728ab076" containerID="08a0e6ede18b68a75633e43e879730e710eb7751a34a96b8e20afe8d47355b32" exitCode=0 Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.110954 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-74mh7" event={"ID":"85470b07-1f66-4a8a-b30c-85c2728ab076","Type":"ContainerDied","Data":"08a0e6ede18b68a75633e43e879730e710eb7751a34a96b8e20afe8d47355b32"} Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.119888 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"cecddbebfa11e45c5f1dd0097b4c03899be17b4acf5853e276d2670a07ce8b39"} Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.119952 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c78941df62e68ce2d8789e0882ac55e8e5e5e1b6be99327c296cf30f5e7acc1a"} Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.120305 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.146498 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94gnq\" (UniqueName: \"kubernetes.io/projected/eaca0644-a305-41b3-84f1-4179e3412d37-kube-api-access-94gnq\") pod \"redhat-operators-hwdhx\" (UID: \"eaca0644-a305-41b3-84f1-4179e3412d37\") " pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.146581 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaca0644-a305-41b3-84f1-4179e3412d37-catalog-content\") pod \"redhat-operators-hwdhx\" (UID: \"eaca0644-a305-41b3-84f1-4179e3412d37\") " pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.146610 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaca0644-a305-41b3-84f1-4179e3412d37-utilities\") pod \"redhat-operators-hwdhx\" (UID: \"eaca0644-a305-41b3-84f1-4179e3412d37\") " pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.148637 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaca0644-a305-41b3-84f1-4179e3412d37-catalog-content\") pod \"redhat-operators-hwdhx\" (UID: \"eaca0644-a305-41b3-84f1-4179e3412d37\") " pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.155406 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaca0644-a305-41b3-84f1-4179e3412d37-utilities\") pod \"redhat-operators-hwdhx\" (UID: \"eaca0644-a305-41b3-84f1-4179e3412d37\") " pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.186167 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94gnq\" (UniqueName: \"kubernetes.io/projected/eaca0644-a305-41b3-84f1-4179e3412d37-kube-api-access-94gnq\") pod \"redhat-operators-hwdhx\" (UID: \"eaca0644-a305-41b3-84f1-4179e3412d37\") " pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.214484 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c2d5j\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.299265 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.346618 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.351167 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjp4q"] Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.375805 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-74cvd"] Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.379403 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.397468 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-74cvd"] Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.565767 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e792b62d-a7ed-4c50-8d30-90bb6b11b367-catalog-content\") pod \"redhat-operators-74cvd\" (UID: \"e792b62d-a7ed-4c50-8d30-90bb6b11b367\") " pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.565828 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brkj8\" (UniqueName: \"kubernetes.io/projected/e792b62d-a7ed-4c50-8d30-90bb6b11b367-kube-api-access-brkj8\") pod \"redhat-operators-74cvd\" (UID: \"e792b62d-a7ed-4c50-8d30-90bb6b11b367\") " pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.565889 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e792b62d-a7ed-4c50-8d30-90bb6b11b367-utilities\") pod \"redhat-operators-74cvd\" (UID: \"e792b62d-a7ed-4c50-8d30-90bb6b11b367\") " pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.624683 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.667329 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4f1c998-fb99-4827-a353-d8b59a309cee-config-volume\") pod \"b4f1c998-fb99-4827-a353-d8b59a309cee\" (UID: \"b4f1c998-fb99-4827-a353-d8b59a309cee\") " Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.667479 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4f1c998-fb99-4827-a353-d8b59a309cee-secret-volume\") pod \"b4f1c998-fb99-4827-a353-d8b59a309cee\" (UID: \"b4f1c998-fb99-4827-a353-d8b59a309cee\") " Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.667532 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvsr6\" (UniqueName: \"kubernetes.io/projected/b4f1c998-fb99-4827-a353-d8b59a309cee-kube-api-access-kvsr6\") pod \"b4f1c998-fb99-4827-a353-d8b59a309cee\" (UID: \"b4f1c998-fb99-4827-a353-d8b59a309cee\") " Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.667693 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e792b62d-a7ed-4c50-8d30-90bb6b11b367-utilities\") pod \"redhat-operators-74cvd\" (UID: \"e792b62d-a7ed-4c50-8d30-90bb6b11b367\") " pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.667766 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e792b62d-a7ed-4c50-8d30-90bb6b11b367-catalog-content\") pod \"redhat-operators-74cvd\" (UID: \"e792b62d-a7ed-4c50-8d30-90bb6b11b367\") " pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.667792 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brkj8\" (UniqueName: \"kubernetes.io/projected/e792b62d-a7ed-4c50-8d30-90bb6b11b367-kube-api-access-brkj8\") pod \"redhat-operators-74cvd\" (UID: \"e792b62d-a7ed-4c50-8d30-90bb6b11b367\") " pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.668914 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4f1c998-fb99-4827-a353-d8b59a309cee-config-volume" (OuterVolumeSpecName: "config-volume") pod "b4f1c998-fb99-4827-a353-d8b59a309cee" (UID: "b4f1c998-fb99-4827-a353-d8b59a309cee"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.670152 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e792b62d-a7ed-4c50-8d30-90bb6b11b367-utilities\") pod \"redhat-operators-74cvd\" (UID: \"e792b62d-a7ed-4c50-8d30-90bb6b11b367\") " pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.673077 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e792b62d-a7ed-4c50-8d30-90bb6b11b367-catalog-content\") pod \"redhat-operators-74cvd\" (UID: \"e792b62d-a7ed-4c50-8d30-90bb6b11b367\") " pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.690812 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4f1c998-fb99-4827-a353-d8b59a309cee-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b4f1c998-fb99-4827-a353-d8b59a309cee" (UID: "b4f1c998-fb99-4827-a353-d8b59a309cee"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.691476 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4f1c998-fb99-4827-a353-d8b59a309cee-kube-api-access-kvsr6" (OuterVolumeSpecName: "kube-api-access-kvsr6") pod "b4f1c998-fb99-4827-a353-d8b59a309cee" (UID: "b4f1c998-fb99-4827-a353-d8b59a309cee"). InnerVolumeSpecName "kube-api-access-kvsr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.693681 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brkj8\" (UniqueName: \"kubernetes.io/projected/e792b62d-a7ed-4c50-8d30-90bb6b11b367-kube-api-access-brkj8\") pod \"redhat-operators-74cvd\" (UID: \"e792b62d-a7ed-4c50-8d30-90bb6b11b367\") " pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.720115 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.746682 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:14 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:14 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:14 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.746731 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.769077 4806 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b4f1c998-fb99-4827-a353-d8b59a309cee-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.769113 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvsr6\" (UniqueName: \"kubernetes.io/projected/b4f1c998-fb99-4827-a353-d8b59a309cee-kube-api-access-kvsr6\") on node \"crc\" DevicePath \"\"" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.769125 4806 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b4f1c998-fb99-4827-a353-d8b59a309cee-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 03:57:14 crc kubenswrapper[4806]: I1204 03:57:14.853708 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c2d5j"] Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.031757 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hwdhx"] Dec 04 03:57:15 crc kubenswrapper[4806]: W1204 03:57:15.059601 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeaca0644_a305_41b3_84f1_4179e3412d37.slice/crio-43424386062cf65cfae41f39a858e3613cdd4b0aaac1d9df3fb05ea8f3f1c555 WatchSource:0}: Error finding container 43424386062cf65cfae41f39a858e3613cdd4b0aaac1d9df3fb05ea8f3f1c555: Status 404 returned error can't find the container with id 43424386062cf65cfae41f39a858e3613cdd4b0aaac1d9df3fb05ea8f3f1c555 Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.086149 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 03:57:15 crc kubenswrapper[4806]: E1204 03:57:15.086673 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4f1c998-fb99-4827-a353-d8b59a309cee" containerName="collect-profiles" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.086694 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4f1c998-fb99-4827-a353-d8b59a309cee" containerName="collect-profiles" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.086836 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4f1c998-fb99-4827-a353-d8b59a309cee" containerName="collect-profiles" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.087579 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.091463 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.091632 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.111182 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.159638 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" event={"ID":"233ad320-ca2f-40bf-bdcf-64e5aed5509f","Type":"ContainerStarted","Data":"e6d5c3fe8047b8d92915969a5f24a5965881126ac40f73d758137e4f2d95b3fd"} Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.183143 4806 generic.go:334] "Generic (PLEG): container finished" podID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" containerID="20071198a89a3299430557f0dcb1a2ae36ec656aaf903936cfdea7dcb15a09ed" exitCode=0 Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.183541 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ln2r5" event={"ID":"ef7e1243-9a57-4bbc-a28a-43cb93cf2add","Type":"ContainerDied","Data":"20071198a89a3299430557f0dcb1a2ae36ec656aaf903936cfdea7dcb15a09ed"} Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.183613 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ln2r5" event={"ID":"ef7e1243-9a57-4bbc-a28a-43cb93cf2add","Type":"ContainerStarted","Data":"99116f9e4ba44a8de33bc4f00aaf26a6059a9a5d23db206cc189bc6b343fec9d"} Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.191669 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwdhx" event={"ID":"eaca0644-a305-41b3-84f1-4179e3412d37","Type":"ContainerStarted","Data":"43424386062cf65cfae41f39a858e3613cdd4b0aaac1d9df3fb05ea8f3f1c555"} Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.194128 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-74cvd"] Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.195454 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" event={"ID":"b4f1c998-fb99-4827-a353-d8b59a309cee","Type":"ContainerDied","Data":"f47d90d516eca9a8057f5d5b24578c8ee9b5c880ba57bfe4167c45af570877ce"} Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.195496 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f47d90d516eca9a8057f5d5b24578c8ee9b5c880ba57bfe4167c45af570877ce" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.195563 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.214879 4806 generic.go:334] "Generic (PLEG): container finished" podID="8d1951d0-8637-4110-ba57-ac3ee5f6c378" containerID="b88f4c91a83516dd76144063318ec60ea57a4f5ced9762e29a6af263a0c67cba" exitCode=0 Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.215095 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjp4q" event={"ID":"8d1951d0-8637-4110-ba57-ac3ee5f6c378","Type":"ContainerDied","Data":"b88f4c91a83516dd76144063318ec60ea57a4f5ced9762e29a6af263a0c67cba"} Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.215126 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjp4q" event={"ID":"8d1951d0-8637-4110-ba57-ac3ee5f6c378","Type":"ContainerStarted","Data":"cf3a45ee2a81d2c903d4a27b28deebe8deacba41da9e4a81e4f1a75abfb8a466"} Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.276517 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4e8bc561-c5a0-4920-9ab3-b852eaa7732f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4e8bc561-c5a0-4920-9ab3-b852eaa7732f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.276589 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4e8bc561-c5a0-4920-9ab3-b852eaa7732f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4e8bc561-c5a0-4920-9ab3-b852eaa7732f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.378339 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4e8bc561-c5a0-4920-9ab3-b852eaa7732f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4e8bc561-c5a0-4920-9ab3-b852eaa7732f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.378723 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4e8bc561-c5a0-4920-9ab3-b852eaa7732f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4e8bc561-c5a0-4920-9ab3-b852eaa7732f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.379135 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4e8bc561-c5a0-4920-9ab3-b852eaa7732f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4e8bc561-c5a0-4920-9ab3-b852eaa7732f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.397836 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4e8bc561-c5a0-4920-9ab3-b852eaa7732f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4e8bc561-c5a0-4920-9ab3-b852eaa7732f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.470603 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.471506 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.746734 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:15 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:15 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:15 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.747064 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.848853 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 03:57:15 crc kubenswrapper[4806]: W1204 03:57:15.921765 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod4e8bc561_c5a0_4920_9ab3_b852eaa7732f.slice/crio-e0c27f1a891f450f081189d49b2c16b7781d9ea5916db74a01fcfb7cbd1f2a53 WatchSource:0}: Error finding container e0c27f1a891f450f081189d49b2c16b7781d9ea5916db74a01fcfb7cbd1f2a53: Status 404 returned error can't find the container with id e0c27f1a891f450f081189d49b2c16b7781d9ea5916db74a01fcfb7cbd1f2a53 Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.924138 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.924993 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.928141 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.936443 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 04 03:57:15 crc kubenswrapper[4806]: I1204 03:57:15.936824 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.096157 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b0ee55d-c4c6-4398-a83d-a3622e6d9442-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4b0ee55d-c4c6-4398-a83d-a3622e6d9442\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.096240 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b0ee55d-c4c6-4398-a83d-a3622e6d9442-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4b0ee55d-c4c6-4398-a83d-a3622e6d9442\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.197618 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b0ee55d-c4c6-4398-a83d-a3622e6d9442-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4b0ee55d-c4c6-4398-a83d-a3622e6d9442\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.197735 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b0ee55d-c4c6-4398-a83d-a3622e6d9442-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4b0ee55d-c4c6-4398-a83d-a3622e6d9442\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.197728 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b0ee55d-c4c6-4398-a83d-a3622e6d9442-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4b0ee55d-c4c6-4398-a83d-a3622e6d9442\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.223537 4806 generic.go:334] "Generic (PLEG): container finished" podID="e792b62d-a7ed-4c50-8d30-90bb6b11b367" containerID="27c810f908ab717540af22e500b0604e032f6208bafceb77e0bdf1eb2e052fa3" exitCode=0 Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.223608 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74cvd" event={"ID":"e792b62d-a7ed-4c50-8d30-90bb6b11b367","Type":"ContainerDied","Data":"27c810f908ab717540af22e500b0604e032f6208bafceb77e0bdf1eb2e052fa3"} Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.223753 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74cvd" event={"ID":"e792b62d-a7ed-4c50-8d30-90bb6b11b367","Type":"ContainerStarted","Data":"b5d87e0a1aacdf3b70102f65783d09327280d6ed70c0416e59a1589ae463dda5"} Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.232082 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b0ee55d-c4c6-4398-a83d-a3622e6d9442-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4b0ee55d-c4c6-4398-a83d-a3622e6d9442\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.249168 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.254261 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" event={"ID":"233ad320-ca2f-40bf-bdcf-64e5aed5509f","Type":"ContainerStarted","Data":"2501b5edc73f99907bc36c581f839bfc572d15b6f6a8af53682d41fdf18d85c9"} Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.254340 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.276612 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4e8bc561-c5a0-4920-9ab3-b852eaa7732f","Type":"ContainerStarted","Data":"e0c27f1a891f450f081189d49b2c16b7781d9ea5916db74a01fcfb7cbd1f2a53"} Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.303807 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" podStartSLOduration=131.303788537 podStartE2EDuration="2m11.303788537s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:16.301980064 +0000 UTC m=+151.160492992" watchObservedRunningTime="2025-12-04 03:57:16.303788537 +0000 UTC m=+151.162301465" Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.343854 4806 generic.go:334] "Generic (PLEG): container finished" podID="eaca0644-a305-41b3-84f1-4179e3412d37" containerID="49e2a09d30e3e8f20dcfef1dd2669937c1f5416d41099e3686311653e8707fa3" exitCode=0 Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.344907 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwdhx" event={"ID":"eaca0644-a305-41b3-84f1-4179e3412d37","Type":"ContainerDied","Data":"49e2a09d30e3e8f20dcfef1dd2669937c1f5416d41099e3686311653e8707fa3"} Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.609176 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.624410 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-7lbq7" Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.761189 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:16 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:16 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:16 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.761252 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:16 crc kubenswrapper[4806]: I1204 03:57:16.870373 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 03:57:17 crc kubenswrapper[4806]: I1204 03:57:17.354964 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4b0ee55d-c4c6-4398-a83d-a3622e6d9442","Type":"ContainerStarted","Data":"c2f97df2c54d45101cca07cdfea22758d7b97c130e7ab78c553042cab55459a5"} Dec 04 03:57:17 crc kubenswrapper[4806]: I1204 03:57:17.360564 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4e8bc561-c5a0-4920-9ab3-b852eaa7732f","Type":"ContainerStarted","Data":"2a1033fc560af5f6cb01d9b3932dbe26b0d58c35bfea896423e7eb9911ad0a15"} Dec 04 03:57:17 crc kubenswrapper[4806]: I1204 03:57:17.744712 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:17 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:17 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:17 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:17 crc kubenswrapper[4806]: I1204 03:57:17.744795 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:17 crc kubenswrapper[4806]: I1204 03:57:17.965017 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-f47vx" Dec 04 03:57:18 crc kubenswrapper[4806]: I1204 03:57:18.392794 4806 generic.go:334] "Generic (PLEG): container finished" podID="4e8bc561-c5a0-4920-9ab3-b852eaa7732f" containerID="2a1033fc560af5f6cb01d9b3932dbe26b0d58c35bfea896423e7eb9911ad0a15" exitCode=0 Dec 04 03:57:18 crc kubenswrapper[4806]: I1204 03:57:18.393163 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4e8bc561-c5a0-4920-9ab3-b852eaa7732f","Type":"ContainerDied","Data":"2a1033fc560af5f6cb01d9b3932dbe26b0d58c35bfea896423e7eb9911ad0a15"} Dec 04 03:57:18 crc kubenswrapper[4806]: I1204 03:57:18.400544 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4b0ee55d-c4c6-4398-a83d-a3622e6d9442","Type":"ContainerStarted","Data":"9dff98bb8c076687f1888c5c587039b3fd384b53a8f5dd128857c26103794cb8"} Dec 04 03:57:18 crc kubenswrapper[4806]: I1204 03:57:18.432705 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.432684703 podStartE2EDuration="3.432684703s" podCreationTimestamp="2025-12-04 03:57:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:18.430447045 +0000 UTC m=+153.288959963" watchObservedRunningTime="2025-12-04 03:57:18.432684703 +0000 UTC m=+153.291197631" Dec 04 03:57:18 crc kubenswrapper[4806]: I1204 03:57:18.744666 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:18 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:18 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:18 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:18 crc kubenswrapper[4806]: I1204 03:57:18.744750 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:19 crc kubenswrapper[4806]: I1204 03:57:19.472554 4806 generic.go:334] "Generic (PLEG): container finished" podID="4b0ee55d-c4c6-4398-a83d-a3622e6d9442" containerID="9dff98bb8c076687f1888c5c587039b3fd384b53a8f5dd128857c26103794cb8" exitCode=0 Dec 04 03:57:19 crc kubenswrapper[4806]: I1204 03:57:19.474392 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4b0ee55d-c4c6-4398-a83d-a3622e6d9442","Type":"ContainerDied","Data":"9dff98bb8c076687f1888c5c587039b3fd384b53a8f5dd128857c26103794cb8"} Dec 04 03:57:19 crc kubenswrapper[4806]: I1204 03:57:19.776137 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:19 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:19 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:19 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:19 crc kubenswrapper[4806]: I1204 03:57:19.776200 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:19 crc kubenswrapper[4806]: I1204 03:57:19.891901 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 03:57:19 crc kubenswrapper[4806]: I1204 03:57:19.963827 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4e8bc561-c5a0-4920-9ab3-b852eaa7732f-kube-api-access\") pod \"4e8bc561-c5a0-4920-9ab3-b852eaa7732f\" (UID: \"4e8bc561-c5a0-4920-9ab3-b852eaa7732f\") " Dec 04 03:57:19 crc kubenswrapper[4806]: I1204 03:57:19.966153 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4e8bc561-c5a0-4920-9ab3-b852eaa7732f-kubelet-dir\") pod \"4e8bc561-c5a0-4920-9ab3-b852eaa7732f\" (UID: \"4e8bc561-c5a0-4920-9ab3-b852eaa7732f\") " Dec 04 03:57:19 crc kubenswrapper[4806]: I1204 03:57:19.966406 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4e8bc561-c5a0-4920-9ab3-b852eaa7732f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4e8bc561-c5a0-4920-9ab3-b852eaa7732f" (UID: "4e8bc561-c5a0-4920-9ab3-b852eaa7732f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 03:57:19 crc kubenswrapper[4806]: I1204 03:57:19.966790 4806 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4e8bc561-c5a0-4920-9ab3-b852eaa7732f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 03:57:19 crc kubenswrapper[4806]: I1204 03:57:19.969835 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e8bc561-c5a0-4920-9ab3-b852eaa7732f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4e8bc561-c5a0-4920-9ab3-b852eaa7732f" (UID: "4e8bc561-c5a0-4920-9ab3-b852eaa7732f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:57:20 crc kubenswrapper[4806]: I1204 03:57:20.068382 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4e8bc561-c5a0-4920-9ab3-b852eaa7732f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 03:57:20 crc kubenswrapper[4806]: I1204 03:57:20.495142 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 03:57:20 crc kubenswrapper[4806]: I1204 03:57:20.495126 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4e8bc561-c5a0-4920-9ab3-b852eaa7732f","Type":"ContainerDied","Data":"e0c27f1a891f450f081189d49b2c16b7781d9ea5916db74a01fcfb7cbd1f2a53"} Dec 04 03:57:20 crc kubenswrapper[4806]: I1204 03:57:20.495277 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0c27f1a891f450f081189d49b2c16b7781d9ea5916db74a01fcfb7cbd1f2a53" Dec 04 03:57:20 crc kubenswrapper[4806]: I1204 03:57:20.744859 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:20 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:20 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:20 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:20 crc kubenswrapper[4806]: I1204 03:57:20.744920 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:20 crc kubenswrapper[4806]: I1204 03:57:20.846571 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 03:57:20 crc kubenswrapper[4806]: I1204 03:57:20.991728 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b0ee55d-c4c6-4398-a83d-a3622e6d9442-kube-api-access\") pod \"4b0ee55d-c4c6-4398-a83d-a3622e6d9442\" (UID: \"4b0ee55d-c4c6-4398-a83d-a3622e6d9442\") " Dec 04 03:57:20 crc kubenswrapper[4806]: I1204 03:57:20.991861 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b0ee55d-c4c6-4398-a83d-a3622e6d9442-kubelet-dir\") pod \"4b0ee55d-c4c6-4398-a83d-a3622e6d9442\" (UID: \"4b0ee55d-c4c6-4398-a83d-a3622e6d9442\") " Dec 04 03:57:20 crc kubenswrapper[4806]: I1204 03:57:20.992178 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b0ee55d-c4c6-4398-a83d-a3622e6d9442-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4b0ee55d-c4c6-4398-a83d-a3622e6d9442" (UID: "4b0ee55d-c4c6-4398-a83d-a3622e6d9442"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 03:57:20 crc kubenswrapper[4806]: I1204 03:57:20.997546 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b0ee55d-c4c6-4398-a83d-a3622e6d9442-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4b0ee55d-c4c6-4398-a83d-a3622e6d9442" (UID: "4b0ee55d-c4c6-4398-a83d-a3622e6d9442"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:57:21 crc kubenswrapper[4806]: I1204 03:57:21.093753 4806 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b0ee55d-c4c6-4398-a83d-a3622e6d9442-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 03:57:21 crc kubenswrapper[4806]: I1204 03:57:21.093800 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b0ee55d-c4c6-4398-a83d-a3622e6d9442-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 03:57:21 crc kubenswrapper[4806]: I1204 03:57:21.520823 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4b0ee55d-c4c6-4398-a83d-a3622e6d9442","Type":"ContainerDied","Data":"c2f97df2c54d45101cca07cdfea22758d7b97c130e7ab78c553042cab55459a5"} Dec 04 03:57:21 crc kubenswrapper[4806]: I1204 03:57:21.521159 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2f97df2c54d45101cca07cdfea22758d7b97c130e7ab78c553042cab55459a5" Dec 04 03:57:21 crc kubenswrapper[4806]: I1204 03:57:21.520872 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 03:57:21 crc kubenswrapper[4806]: I1204 03:57:21.752837 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:21 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:21 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:21 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:21 crc kubenswrapper[4806]: I1204 03:57:21.752970 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:22 crc kubenswrapper[4806]: I1204 03:57:22.405850 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-8r5hx" Dec 04 03:57:22 crc kubenswrapper[4806]: I1204 03:57:22.744404 4806 patch_prober.go:28] interesting pod/router-default-5444994796-mmxrh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 03:57:22 crc kubenswrapper[4806]: [-]has-synced failed: reason withheld Dec 04 03:57:22 crc kubenswrapper[4806]: [+]process-running ok Dec 04 03:57:22 crc kubenswrapper[4806]: healthz check failed Dec 04 03:57:22 crc kubenswrapper[4806]: I1204 03:57:22.744475 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-mmxrh" podUID="861e4822-a809-46c7-af83-06e8978ca8aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 03:57:22 crc kubenswrapper[4806]: I1204 03:57:22.793665 4806 patch_prober.go:28] interesting pod/console-f9d7485db-z2mdj container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Dec 04 03:57:22 crc kubenswrapper[4806]: I1204 03:57:22.793720 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-z2mdj" podUID="a3f377d2-10db-41db-9361-06abff2e0d29" containerName="console" probeResult="failure" output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" Dec 04 03:57:23 crc kubenswrapper[4806]: I1204 03:57:23.745595 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:23 crc kubenswrapper[4806]: I1204 03:57:23.752271 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-mmxrh" Dec 04 03:57:27 crc kubenswrapper[4806]: I1204 03:57:27.047091 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 03:57:27 crc kubenswrapper[4806]: I1204 03:57:27.047504 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 03:57:27 crc kubenswrapper[4806]: I1204 03:57:27.754245 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs\") pod \"network-metrics-daemon-b9khd\" (UID: \"41fdb600-3868-48ab-8396-17c2382b6168\") " pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:57:27 crc kubenswrapper[4806]: I1204 03:57:27.773383 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41fdb600-3868-48ab-8396-17c2382b6168-metrics-certs\") pod \"network-metrics-daemon-b9khd\" (UID: \"41fdb600-3868-48ab-8396-17c2382b6168\") " pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:57:28 crc kubenswrapper[4806]: I1204 03:57:28.070602 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b9khd" Dec 04 03:57:32 crc kubenswrapper[4806]: I1204 03:57:32.797529 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:32 crc kubenswrapper[4806]: I1204 03:57:32.802110 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 03:57:34 crc kubenswrapper[4806]: I1204 03:57:34.351673 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 03:57:42 crc kubenswrapper[4806]: I1204 03:57:42.850772 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5r9n4" Dec 04 03:57:49 crc kubenswrapper[4806]: E1204 03:57:49.654416 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 04 03:57:49 crc kubenswrapper[4806]: E1204 03:57:49.655180 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wf6x7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ln2r5_openshift-marketplace(ef7e1243-9a57-4bbc-a28a-43cb93cf2add): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 03:57:49 crc kubenswrapper[4806]: E1204 03:57:49.656350 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-ln2r5" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" Dec 04 03:57:49 crc kubenswrapper[4806]: E1204 03:57:49.749589 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ln2r5" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" Dec 04 03:57:49 crc kubenswrapper[4806]: E1204 03:57:49.937014 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 04 03:57:49 crc kubenswrapper[4806]: E1204 03:57:49.937546 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h4574,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-zjp4q_openshift-marketplace(8d1951d0-8637-4110-ba57-ac3ee5f6c378): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 03:57:49 crc kubenswrapper[4806]: E1204 03:57:49.938803 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-zjp4q" podUID="8d1951d0-8637-4110-ba57-ac3ee5f6c378" Dec 04 03:57:50 crc kubenswrapper[4806]: I1204 03:57:50.145409 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-b9khd"] Dec 04 03:57:50 crc kubenswrapper[4806]: I1204 03:57:50.752535 4806 generic.go:334] "Generic (PLEG): container finished" podID="85470b07-1f66-4a8a-b30c-85c2728ab076" containerID="497afd2e41a71efe926f63ca1fb6fbfbde3e8bfce08650ea57e99c5b4a96fc8a" exitCode=0 Dec 04 03:57:50 crc kubenswrapper[4806]: I1204 03:57:50.752625 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-74mh7" event={"ID":"85470b07-1f66-4a8a-b30c-85c2728ab076","Type":"ContainerDied","Data":"497afd2e41a71efe926f63ca1fb6fbfbde3e8bfce08650ea57e99c5b4a96fc8a"} Dec 04 03:57:50 crc kubenswrapper[4806]: I1204 03:57:50.762419 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwdhx" event={"ID":"eaca0644-a305-41b3-84f1-4179e3412d37","Type":"ContainerStarted","Data":"32654e96eee457dd1b602189dfdb36283e390b07b7a62a8eae9f24d7b2ce8d7d"} Dec 04 03:57:50 crc kubenswrapper[4806]: I1204 03:57:50.764593 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b9khd" event={"ID":"41fdb600-3868-48ab-8396-17c2382b6168","Type":"ContainerStarted","Data":"2336e253d201ef6b35a9f79f926e14471125e9b8abc546a2af65b9c6e74af069"} Dec 04 03:57:50 crc kubenswrapper[4806]: I1204 03:57:50.767332 4806 generic.go:334] "Generic (PLEG): container finished" podID="07d1500e-9acd-47da-b1cb-13a7c7123f2a" containerID="7a63f9ee49f8b4b935ce20874d60f000009fd9bb7ac7c65a5e2285c70b6412c2" exitCode=0 Dec 04 03:57:50 crc kubenswrapper[4806]: I1204 03:57:50.767408 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brh2b" event={"ID":"07d1500e-9acd-47da-b1cb-13a7c7123f2a","Type":"ContainerDied","Data":"7a63f9ee49f8b4b935ce20874d60f000009fd9bb7ac7c65a5e2285c70b6412c2"} Dec 04 03:57:50 crc kubenswrapper[4806]: I1204 03:57:50.769270 4806 generic.go:334] "Generic (PLEG): container finished" podID="59a2e27f-ed83-42ba-9371-14546fbcc6af" containerID="769dff3547b6a3de75992d668dcbd705b51700b24d99cf13cb7499c2124acb1b" exitCode=0 Dec 04 03:57:50 crc kubenswrapper[4806]: I1204 03:57:50.769322 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q9fqz" event={"ID":"59a2e27f-ed83-42ba-9371-14546fbcc6af","Type":"ContainerDied","Data":"769dff3547b6a3de75992d668dcbd705b51700b24d99cf13cb7499c2124acb1b"} Dec 04 03:57:50 crc kubenswrapper[4806]: I1204 03:57:50.781218 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74cvd" event={"ID":"e792b62d-a7ed-4c50-8d30-90bb6b11b367","Type":"ContainerStarted","Data":"47927e658b525cfd6f895efb0a00d03af4078239d7c763e31c023a7a41a6dbe3"} Dec 04 03:57:50 crc kubenswrapper[4806]: I1204 03:57:50.783383 4806 generic.go:334] "Generic (PLEG): container finished" podID="21c59133-1783-4225-a70d-1395d9642980" containerID="27607f453517689111573accd547468b759759e2db407a8d687a6b85a5eee15d" exitCode=0 Dec 04 03:57:50 crc kubenswrapper[4806]: I1204 03:57:50.783425 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlrnj" event={"ID":"21c59133-1783-4225-a70d-1395d9642980","Type":"ContainerDied","Data":"27607f453517689111573accd547468b759759e2db407a8d687a6b85a5eee15d"} Dec 04 03:57:50 crc kubenswrapper[4806]: I1204 03:57:50.937528 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7m42k"] Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.561532 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.662835 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 03:57:51 crc kubenswrapper[4806]: E1204 03:57:51.663271 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e8bc561-c5a0-4920-9ab3-b852eaa7732f" containerName="pruner" Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.663297 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e8bc561-c5a0-4920-9ab3-b852eaa7732f" containerName="pruner" Dec 04 03:57:51 crc kubenswrapper[4806]: E1204 03:57:51.663309 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b0ee55d-c4c6-4398-a83d-a3622e6d9442" containerName="pruner" Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.663317 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b0ee55d-c4c6-4398-a83d-a3622e6d9442" containerName="pruner" Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.663436 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b0ee55d-c4c6-4398-a83d-a3622e6d9442" containerName="pruner" Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.663457 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e8bc561-c5a0-4920-9ab3-b852eaa7732f" containerName="pruner" Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.663883 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.672380 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.673139 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.676886 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f8abe4ca-99e2-4390-b893-c69229d8f3c5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f8abe4ca-99e2-4390-b893-c69229d8f3c5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.677004 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f8abe4ca-99e2-4390-b893-c69229d8f3c5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f8abe4ca-99e2-4390-b893-c69229d8f3c5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.682762 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.778744 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f8abe4ca-99e2-4390-b893-c69229d8f3c5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f8abe4ca-99e2-4390-b893-c69229d8f3c5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.778818 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f8abe4ca-99e2-4390-b893-c69229d8f3c5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f8abe4ca-99e2-4390-b893-c69229d8f3c5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.778897 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f8abe4ca-99e2-4390-b893-c69229d8f3c5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f8abe4ca-99e2-4390-b893-c69229d8f3c5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.808649 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b9khd" event={"ID":"41fdb600-3868-48ab-8396-17c2382b6168","Type":"ContainerStarted","Data":"099e533338d8b6a8b1951dee33330651d3b74fddf5edd05dd6f6a91cb3058e74"} Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.815482 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f8abe4ca-99e2-4390-b893-c69229d8f3c5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f8abe4ca-99e2-4390-b893-c69229d8f3c5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 03:57:51 crc kubenswrapper[4806]: I1204 03:57:51.984539 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 03:57:52 crc kubenswrapper[4806]: I1204 03:57:52.388807 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 03:57:52 crc kubenswrapper[4806]: I1204 03:57:52.828520 4806 generic.go:334] "Generic (PLEG): container finished" podID="eaca0644-a305-41b3-84f1-4179e3412d37" containerID="32654e96eee457dd1b602189dfdb36283e390b07b7a62a8eae9f24d7b2ce8d7d" exitCode=0 Dec 04 03:57:52 crc kubenswrapper[4806]: I1204 03:57:52.829022 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwdhx" event={"ID":"eaca0644-a305-41b3-84f1-4179e3412d37","Type":"ContainerDied","Data":"32654e96eee457dd1b602189dfdb36283e390b07b7a62a8eae9f24d7b2ce8d7d"} Dec 04 03:57:52 crc kubenswrapper[4806]: I1204 03:57:52.834260 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b9khd" event={"ID":"41fdb600-3868-48ab-8396-17c2382b6168","Type":"ContainerStarted","Data":"3f5ee979a7433d7c415c748f078ec1db245c79d58b8c8058c548e86043e68dac"} Dec 04 03:57:52 crc kubenswrapper[4806]: I1204 03:57:52.835848 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"f8abe4ca-99e2-4390-b893-c69229d8f3c5","Type":"ContainerStarted","Data":"bb49f0de03b0558ab29aa176ecf23434d2e588b813f14ead232440fa4fa977ae"} Dec 04 03:57:52 crc kubenswrapper[4806]: I1204 03:57:52.841696 4806 generic.go:334] "Generic (PLEG): container finished" podID="e792b62d-a7ed-4c50-8d30-90bb6b11b367" containerID="47927e658b525cfd6f895efb0a00d03af4078239d7c763e31c023a7a41a6dbe3" exitCode=0 Dec 04 03:57:52 crc kubenswrapper[4806]: I1204 03:57:52.841730 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74cvd" event={"ID":"e792b62d-a7ed-4c50-8d30-90bb6b11b367","Type":"ContainerDied","Data":"47927e658b525cfd6f895efb0a00d03af4078239d7c763e31c023a7a41a6dbe3"} Dec 04 03:57:52 crc kubenswrapper[4806]: I1204 03:57:52.888699 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-b9khd" podStartSLOduration=167.888679324 podStartE2EDuration="2m47.888679324s" podCreationTimestamp="2025-12-04 03:55:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:52.887394869 +0000 UTC m=+187.745907797" watchObservedRunningTime="2025-12-04 03:57:52.888679324 +0000 UTC m=+187.747192252" Dec 04 03:57:54 crc kubenswrapper[4806]: I1204 03:57:54.857638 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"f8abe4ca-99e2-4390-b893-c69229d8f3c5","Type":"ContainerStarted","Data":"f9996bfe9e09b4872f0d2171932a4e08f076bdfe8b47a2e25758c3003f3707e4"} Dec 04 03:57:54 crc kubenswrapper[4806]: I1204 03:57:54.864862 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-74mh7" event={"ID":"85470b07-1f66-4a8a-b30c-85c2728ab076","Type":"ContainerStarted","Data":"aca9fad4657b3ae5075c8e5de91680863a86206470c07073440a32a530c37d5e"} Dec 04 03:57:54 crc kubenswrapper[4806]: I1204 03:57:54.870762 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=3.87074057 podStartE2EDuration="3.87074057s" podCreationTimestamp="2025-12-04 03:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:57:54.868884485 +0000 UTC m=+189.727397423" watchObservedRunningTime="2025-12-04 03:57:54.87074057 +0000 UTC m=+189.729253498" Dec 04 03:57:54 crc kubenswrapper[4806]: I1204 03:57:54.888884 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-74mh7" podStartSLOduration=4.574602411 podStartE2EDuration="43.888865318s" podCreationTimestamp="2025-12-04 03:57:11 +0000 UTC" firstStartedPulling="2025-12-04 03:57:14.113008676 +0000 UTC m=+148.971521604" lastFinishedPulling="2025-12-04 03:57:53.427271583 +0000 UTC m=+188.285784511" observedRunningTime="2025-12-04 03:57:54.887691387 +0000 UTC m=+189.746204315" watchObservedRunningTime="2025-12-04 03:57:54.888865318 +0000 UTC m=+189.747378246" Dec 04 03:57:55 crc kubenswrapper[4806]: I1204 03:57:55.871508 4806 generic.go:334] "Generic (PLEG): container finished" podID="f8abe4ca-99e2-4390-b893-c69229d8f3c5" containerID="f9996bfe9e09b4872f0d2171932a4e08f076bdfe8b47a2e25758c3003f3707e4" exitCode=0 Dec 04 03:57:55 crc kubenswrapper[4806]: I1204 03:57:55.871578 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"f8abe4ca-99e2-4390-b893-c69229d8f3c5","Type":"ContainerDied","Data":"f9996bfe9e09b4872f0d2171932a4e08f076bdfe8b47a2e25758c3003f3707e4"} Dec 04 03:57:57 crc kubenswrapper[4806]: I1204 03:57:57.047048 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 03:57:57 crc kubenswrapper[4806]: I1204 03:57:57.047105 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 03:57:57 crc kubenswrapper[4806]: I1204 03:57:57.133247 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 03:57:57 crc kubenswrapper[4806]: I1204 03:57:57.278444 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f8abe4ca-99e2-4390-b893-c69229d8f3c5-kube-api-access\") pod \"f8abe4ca-99e2-4390-b893-c69229d8f3c5\" (UID: \"f8abe4ca-99e2-4390-b893-c69229d8f3c5\") " Dec 04 03:57:57 crc kubenswrapper[4806]: I1204 03:57:57.278604 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f8abe4ca-99e2-4390-b893-c69229d8f3c5-kubelet-dir\") pod \"f8abe4ca-99e2-4390-b893-c69229d8f3c5\" (UID: \"f8abe4ca-99e2-4390-b893-c69229d8f3c5\") " Dec 04 03:57:57 crc kubenswrapper[4806]: I1204 03:57:57.278848 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8abe4ca-99e2-4390-b893-c69229d8f3c5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f8abe4ca-99e2-4390-b893-c69229d8f3c5" (UID: "f8abe4ca-99e2-4390-b893-c69229d8f3c5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 03:57:57 crc kubenswrapper[4806]: I1204 03:57:57.302157 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8abe4ca-99e2-4390-b893-c69229d8f3c5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f8abe4ca-99e2-4390-b893-c69229d8f3c5" (UID: "f8abe4ca-99e2-4390-b893-c69229d8f3c5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:57:57 crc kubenswrapper[4806]: I1204 03:57:57.382340 4806 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f8abe4ca-99e2-4390-b893-c69229d8f3c5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 03:57:57 crc kubenswrapper[4806]: I1204 03:57:57.382383 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f8abe4ca-99e2-4390-b893-c69229d8f3c5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 03:57:57 crc kubenswrapper[4806]: I1204 03:57:57.881870 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"f8abe4ca-99e2-4390-b893-c69229d8f3c5","Type":"ContainerDied","Data":"bb49f0de03b0558ab29aa176ecf23434d2e588b813f14ead232440fa4fa977ae"} Dec 04 03:57:57 crc kubenswrapper[4806]: I1204 03:57:57.881977 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb49f0de03b0558ab29aa176ecf23434d2e588b813f14ead232440fa4fa977ae" Dec 04 03:57:57 crc kubenswrapper[4806]: I1204 03:57:57.881992 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.658759 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 03:57:58 crc kubenswrapper[4806]: E1204 03:57:58.659354 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8abe4ca-99e2-4390-b893-c69229d8f3c5" containerName="pruner" Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.659370 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8abe4ca-99e2-4390-b893-c69229d8f3c5" containerName="pruner" Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.659503 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8abe4ca-99e2-4390-b893-c69229d8f3c5" containerName="pruner" Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.660173 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.662088 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.662242 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.671889 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.798616 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed2b244e-28d2-406e-8b09-68e91494fa95-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ed2b244e-28d2-406e-8b09-68e91494fa95\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.798709 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed2b244e-28d2-406e-8b09-68e91494fa95-kube-api-access\") pod \"installer-9-crc\" (UID: \"ed2b244e-28d2-406e-8b09-68e91494fa95\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.798747 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ed2b244e-28d2-406e-8b09-68e91494fa95-var-lock\") pod \"installer-9-crc\" (UID: \"ed2b244e-28d2-406e-8b09-68e91494fa95\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.899448 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed2b244e-28d2-406e-8b09-68e91494fa95-kube-api-access\") pod \"installer-9-crc\" (UID: \"ed2b244e-28d2-406e-8b09-68e91494fa95\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.899514 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ed2b244e-28d2-406e-8b09-68e91494fa95-var-lock\") pod \"installer-9-crc\" (UID: \"ed2b244e-28d2-406e-8b09-68e91494fa95\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.899568 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed2b244e-28d2-406e-8b09-68e91494fa95-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ed2b244e-28d2-406e-8b09-68e91494fa95\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.899883 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ed2b244e-28d2-406e-8b09-68e91494fa95-var-lock\") pod \"installer-9-crc\" (UID: \"ed2b244e-28d2-406e-8b09-68e91494fa95\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.899905 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed2b244e-28d2-406e-8b09-68e91494fa95-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ed2b244e-28d2-406e-8b09-68e91494fa95\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.918168 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed2b244e-28d2-406e-8b09-68e91494fa95-kube-api-access\") pod \"installer-9-crc\" (UID: \"ed2b244e-28d2-406e-8b09-68e91494fa95\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 03:57:58 crc kubenswrapper[4806]: I1204 03:57:58.979061 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 03:58:00 crc kubenswrapper[4806]: I1204 03:58:00.427477 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 03:58:00 crc kubenswrapper[4806]: W1204 03:58:00.434156 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poded2b244e_28d2_406e_8b09_68e91494fa95.slice/crio-e38cfe438a0976c9472417549ba39b4b16d19bc151f9419dbfe3fee4c7b0cada WatchSource:0}: Error finding container e38cfe438a0976c9472417549ba39b4b16d19bc151f9419dbfe3fee4c7b0cada: Status 404 returned error can't find the container with id e38cfe438a0976c9472417549ba39b4b16d19bc151f9419dbfe3fee4c7b0cada Dec 04 03:58:00 crc kubenswrapper[4806]: I1204 03:58:00.898315 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ed2b244e-28d2-406e-8b09-68e91494fa95","Type":"ContainerStarted","Data":"e38cfe438a0976c9472417549ba39b4b16d19bc151f9419dbfe3fee4c7b0cada"} Dec 04 03:58:01 crc kubenswrapper[4806]: I1204 03:58:01.805237 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:58:01 crc kubenswrapper[4806]: I1204 03:58:01.805641 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:58:01 crc kubenswrapper[4806]: I1204 03:58:01.895435 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:58:01 crc kubenswrapper[4806]: I1204 03:58:01.906157 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brh2b" event={"ID":"07d1500e-9acd-47da-b1cb-13a7c7123f2a","Type":"ContainerStarted","Data":"9d7c7b1f48626d44501f40f7e834749b654026aaf5a15b1c2fa5807facfb542c"} Dec 04 03:58:01 crc kubenswrapper[4806]: I1204 03:58:01.908024 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ed2b244e-28d2-406e-8b09-68e91494fa95","Type":"ContainerStarted","Data":"3982b9a8f94d1548d202054403e6d6a4da1b0311f261c078fe014badffcb4675"} Dec 04 03:58:01 crc kubenswrapper[4806]: I1204 03:58:01.948969 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:58:02 crc kubenswrapper[4806]: I1204 03:58:02.668800 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-74mh7"] Dec 04 03:58:02 crc kubenswrapper[4806]: I1204 03:58:02.915391 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74cvd" event={"ID":"e792b62d-a7ed-4c50-8d30-90bb6b11b367","Type":"ContainerStarted","Data":"b5346e8b5afe8cd994f2b340e662d3c89ea5b253b2fec73f0c47f8c2673da6f3"} Dec 04 03:58:02 crc kubenswrapper[4806]: I1204 03:58:02.917267 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwdhx" event={"ID":"eaca0644-a305-41b3-84f1-4179e3412d37","Type":"ContainerStarted","Data":"aa4c2ceb9165d54155139b13b8e3dd58f12f8aa97ba2236faa3c276ca56eb542"} Dec 04 03:58:02 crc kubenswrapper[4806]: I1204 03:58:02.919892 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q9fqz" event={"ID":"59a2e27f-ed83-42ba-9371-14546fbcc6af","Type":"ContainerStarted","Data":"04dd0126cf689d5f7aa3f5021261660f9566c1317f784ebbc0a0e5b90f9a74df"} Dec 04 03:58:02 crc kubenswrapper[4806]: I1204 03:58:02.922615 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlrnj" event={"ID":"21c59133-1783-4225-a70d-1395d9642980","Type":"ContainerStarted","Data":"84659a3af5d80e162ecbe638130986ac93820569328f6cb1a5f7a30d3d198440"} Dec 04 03:58:02 crc kubenswrapper[4806]: I1204 03:58:02.942113 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-74cvd" podStartSLOduration=5.062355515 podStartE2EDuration="48.942095989s" podCreationTimestamp="2025-12-04 03:57:14 +0000 UTC" firstStartedPulling="2025-12-04 03:57:16.239747898 +0000 UTC m=+151.098260826" lastFinishedPulling="2025-12-04 03:58:00.119488382 +0000 UTC m=+194.978001300" observedRunningTime="2025-12-04 03:58:02.940718958 +0000 UTC m=+197.799231886" watchObservedRunningTime="2025-12-04 03:58:02.942095989 +0000 UTC m=+197.800608917" Dec 04 03:58:02 crc kubenswrapper[4806]: I1204 03:58:02.976624 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rlrnj" podStartSLOduration=8.299093067 podStartE2EDuration="52.976609336s" podCreationTimestamp="2025-12-04 03:57:10 +0000 UTC" firstStartedPulling="2025-12-04 03:57:14.049120631 +0000 UTC m=+148.907633559" lastFinishedPulling="2025-12-04 03:57:58.7266369 +0000 UTC m=+193.585149828" observedRunningTime="2025-12-04 03:58:02.973754391 +0000 UTC m=+197.832267319" watchObservedRunningTime="2025-12-04 03:58:02.976609336 +0000 UTC m=+197.835122264" Dec 04 03:58:02 crc kubenswrapper[4806]: I1204 03:58:02.992822 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hwdhx" podStartSLOduration=6.231505927 podStartE2EDuration="49.992801687s" podCreationTimestamp="2025-12-04 03:57:13 +0000 UTC" firstStartedPulling="2025-12-04 03:57:16.35812232 +0000 UTC m=+151.216635248" lastFinishedPulling="2025-12-04 03:58:00.11941808 +0000 UTC m=+194.977931008" observedRunningTime="2025-12-04 03:58:02.990669227 +0000 UTC m=+197.849182155" watchObservedRunningTime="2025-12-04 03:58:02.992801687 +0000 UTC m=+197.851314615" Dec 04 03:58:03 crc kubenswrapper[4806]: I1204 03:58:03.020770 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q9fqz" podStartSLOduration=6.862296377 podStartE2EDuration="53.020748092s" podCreationTimestamp="2025-12-04 03:57:10 +0000 UTC" firstStartedPulling="2025-12-04 03:57:12.895304465 +0000 UTC m=+147.753817393" lastFinishedPulling="2025-12-04 03:57:59.05375618 +0000 UTC m=+193.912269108" observedRunningTime="2025-12-04 03:58:03.016635819 +0000 UTC m=+197.875148747" watchObservedRunningTime="2025-12-04 03:58:03.020748092 +0000 UTC m=+197.879261020" Dec 04 03:58:03 crc kubenswrapper[4806]: I1204 03:58:03.040383 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=5.040366338 podStartE2EDuration="5.040366338s" podCreationTimestamp="2025-12-04 03:57:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:58:03.035375913 +0000 UTC m=+197.893888841" watchObservedRunningTime="2025-12-04 03:58:03.040366338 +0000 UTC m=+197.898879256" Dec 04 03:58:03 crc kubenswrapper[4806]: I1204 03:58:03.063228 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-brh2b" podStartSLOduration=7.951381524 podStartE2EDuration="52.063210783s" podCreationTimestamp="2025-12-04 03:57:11 +0000 UTC" firstStartedPulling="2025-12-04 03:57:14.04936565 +0000 UTC m=+148.907878578" lastFinishedPulling="2025-12-04 03:57:58.161194899 +0000 UTC m=+193.019707837" observedRunningTime="2025-12-04 03:58:03.058022141 +0000 UTC m=+197.916535069" watchObservedRunningTime="2025-12-04 03:58:03.063210783 +0000 UTC m=+197.921723711" Dec 04 03:58:03 crc kubenswrapper[4806]: I1204 03:58:03.925823 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-74mh7" podUID="85470b07-1f66-4a8a-b30c-85c2728ab076" containerName="registry-server" containerID="cri-o://aca9fad4657b3ae5075c8e5de91680863a86206470c07073440a32a530c37d5e" gracePeriod=2 Dec 04 03:58:04 crc kubenswrapper[4806]: I1204 03:58:04.300463 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:58:04 crc kubenswrapper[4806]: I1204 03:58:04.300837 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:58:04 crc kubenswrapper[4806]: I1204 03:58:04.720709 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:58:04 crc kubenswrapper[4806]: I1204 03:58:04.720762 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:58:04 crc kubenswrapper[4806]: I1204 03:58:04.935271 4806 generic.go:334] "Generic (PLEG): container finished" podID="85470b07-1f66-4a8a-b30c-85c2728ab076" containerID="aca9fad4657b3ae5075c8e5de91680863a86206470c07073440a32a530c37d5e" exitCode=0 Dec 04 03:58:04 crc kubenswrapper[4806]: I1204 03:58:04.935317 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-74mh7" event={"ID":"85470b07-1f66-4a8a-b30c-85c2728ab076","Type":"ContainerDied","Data":"aca9fad4657b3ae5075c8e5de91680863a86206470c07073440a32a530c37d5e"} Dec 04 03:58:05 crc kubenswrapper[4806]: I1204 03:58:05.346521 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hwdhx" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" containerName="registry-server" probeResult="failure" output=< Dec 04 03:58:05 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Dec 04 03:58:05 crc kubenswrapper[4806]: > Dec 04 03:58:05 crc kubenswrapper[4806]: I1204 03:58:05.767424 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-74cvd" podUID="e792b62d-a7ed-4c50-8d30-90bb6b11b367" containerName="registry-server" probeResult="failure" output=< Dec 04 03:58:05 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Dec 04 03:58:05 crc kubenswrapper[4806]: > Dec 04 03:58:06 crc kubenswrapper[4806]: I1204 03:58:06.829380 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:58:06 crc kubenswrapper[4806]: I1204 03:58:06.955511 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-74mh7" event={"ID":"85470b07-1f66-4a8a-b30c-85c2728ab076","Type":"ContainerDied","Data":"5d6a3c5e060d7b05a8b056252ec00c3383e33146bb35cbc6b5ed05263fbc1cad"} Dec 04 03:58:06 crc kubenswrapper[4806]: I1204 03:58:06.955995 4806 scope.go:117] "RemoveContainer" containerID="aca9fad4657b3ae5075c8e5de91680863a86206470c07073440a32a530c37d5e" Dec 04 03:58:06 crc kubenswrapper[4806]: I1204 03:58:06.955832 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-74mh7" Dec 04 03:58:06 crc kubenswrapper[4806]: I1204 03:58:06.979160 4806 scope.go:117] "RemoveContainer" containerID="497afd2e41a71efe926f63ca1fb6fbfbde3e8bfce08650ea57e99c5b4a96fc8a" Dec 04 03:58:06 crc kubenswrapper[4806]: I1204 03:58:06.999171 4806 scope.go:117] "RemoveContainer" containerID="08a0e6ede18b68a75633e43e879730e710eb7751a34a96b8e20afe8d47355b32" Dec 04 03:58:07 crc kubenswrapper[4806]: I1204 03:58:07.025597 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85470b07-1f66-4a8a-b30c-85c2728ab076-catalog-content\") pod \"85470b07-1f66-4a8a-b30c-85c2728ab076\" (UID: \"85470b07-1f66-4a8a-b30c-85c2728ab076\") " Dec 04 03:58:07 crc kubenswrapper[4806]: I1204 03:58:07.025678 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66bms\" (UniqueName: \"kubernetes.io/projected/85470b07-1f66-4a8a-b30c-85c2728ab076-kube-api-access-66bms\") pod \"85470b07-1f66-4a8a-b30c-85c2728ab076\" (UID: \"85470b07-1f66-4a8a-b30c-85c2728ab076\") " Dec 04 03:58:07 crc kubenswrapper[4806]: I1204 03:58:07.025718 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85470b07-1f66-4a8a-b30c-85c2728ab076-utilities\") pod \"85470b07-1f66-4a8a-b30c-85c2728ab076\" (UID: \"85470b07-1f66-4a8a-b30c-85c2728ab076\") " Dec 04 03:58:07 crc kubenswrapper[4806]: I1204 03:58:07.026731 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85470b07-1f66-4a8a-b30c-85c2728ab076-utilities" (OuterVolumeSpecName: "utilities") pod "85470b07-1f66-4a8a-b30c-85c2728ab076" (UID: "85470b07-1f66-4a8a-b30c-85c2728ab076"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:58:07 crc kubenswrapper[4806]: I1204 03:58:07.035166 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85470b07-1f66-4a8a-b30c-85c2728ab076-kube-api-access-66bms" (OuterVolumeSpecName: "kube-api-access-66bms") pod "85470b07-1f66-4a8a-b30c-85c2728ab076" (UID: "85470b07-1f66-4a8a-b30c-85c2728ab076"). InnerVolumeSpecName "kube-api-access-66bms". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:58:07 crc kubenswrapper[4806]: I1204 03:58:07.083151 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85470b07-1f66-4a8a-b30c-85c2728ab076-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85470b07-1f66-4a8a-b30c-85c2728ab076" (UID: "85470b07-1f66-4a8a-b30c-85c2728ab076"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:58:07 crc kubenswrapper[4806]: I1204 03:58:07.126492 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66bms\" (UniqueName: \"kubernetes.io/projected/85470b07-1f66-4a8a-b30c-85c2728ab076-kube-api-access-66bms\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:07 crc kubenswrapper[4806]: I1204 03:58:07.126542 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85470b07-1f66-4a8a-b30c-85c2728ab076-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:07 crc kubenswrapper[4806]: I1204 03:58:07.126555 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85470b07-1f66-4a8a-b30c-85c2728ab076-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:07 crc kubenswrapper[4806]: I1204 03:58:07.282429 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-74mh7"] Dec 04 03:58:07 crc kubenswrapper[4806]: I1204 03:58:07.285647 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-74mh7"] Dec 04 03:58:07 crc kubenswrapper[4806]: I1204 03:58:07.430321 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85470b07-1f66-4a8a-b30c-85c2728ab076" path="/var/lib/kubelet/pods/85470b07-1f66-4a8a-b30c-85c2728ab076/volumes" Dec 04 03:58:07 crc kubenswrapper[4806]: I1204 03:58:07.964211 4806 generic.go:334] "Generic (PLEG): container finished" podID="8d1951d0-8637-4110-ba57-ac3ee5f6c378" containerID="1067ca88719c50540d0bacf11167c7c884852a6f3c739bd3334efb714f20d3f9" exitCode=0 Dec 04 03:58:07 crc kubenswrapper[4806]: I1204 03:58:07.964283 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjp4q" event={"ID":"8d1951d0-8637-4110-ba57-ac3ee5f6c378","Type":"ContainerDied","Data":"1067ca88719c50540d0bacf11167c7c884852a6f3c739bd3334efb714f20d3f9"} Dec 04 03:58:07 crc kubenswrapper[4806]: I1204 03:58:07.968174 4806 generic.go:334] "Generic (PLEG): container finished" podID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" containerID="d9244d156597a2a7cdf05700dbc72398a240c92a12488f2daf0653099b1f4e8d" exitCode=0 Dec 04 03:58:07 crc kubenswrapper[4806]: I1204 03:58:07.968203 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ln2r5" event={"ID":"ef7e1243-9a57-4bbc-a28a-43cb93cf2add","Type":"ContainerDied","Data":"d9244d156597a2a7cdf05700dbc72398a240c92a12488f2daf0653099b1f4e8d"} Dec 04 03:58:08 crc kubenswrapper[4806]: I1204 03:58:08.975775 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjp4q" event={"ID":"8d1951d0-8637-4110-ba57-ac3ee5f6c378","Type":"ContainerStarted","Data":"362f8314a2ffc4b44fbee193dad7a4161eaf444a32d6572e1d6664e60581e82e"} Dec 04 03:58:08 crc kubenswrapper[4806]: I1204 03:58:08.981070 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ln2r5" event={"ID":"ef7e1243-9a57-4bbc-a28a-43cb93cf2add","Type":"ContainerStarted","Data":"3d8a89e3e4eb267dcc1cb982fa7ce09d4b6fe168fad8c21218c0de647968c8f9"} Dec 04 03:58:08 crc kubenswrapper[4806]: I1204 03:58:08.997904 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zjp4q" podStartSLOduration=3.957422622 podStartE2EDuration="55.997886034s" podCreationTimestamp="2025-12-04 03:57:13 +0000 UTC" firstStartedPulling="2025-12-04 03:57:16.358100219 +0000 UTC m=+151.216613147" lastFinishedPulling="2025-12-04 03:58:08.398563601 +0000 UTC m=+203.257076559" observedRunningTime="2025-12-04 03:58:08.994422955 +0000 UTC m=+203.852935883" watchObservedRunningTime="2025-12-04 03:58:08.997886034 +0000 UTC m=+203.856398962" Dec 04 03:58:09 crc kubenswrapper[4806]: I1204 03:58:09.026552 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ln2r5" podStartSLOduration=3.846766659 podStartE2EDuration="57.026528714s" podCreationTimestamp="2025-12-04 03:57:12 +0000 UTC" firstStartedPulling="2025-12-04 03:57:15.201915201 +0000 UTC m=+150.060428129" lastFinishedPulling="2025-12-04 03:58:08.381677256 +0000 UTC m=+203.240190184" observedRunningTime="2025-12-04 03:58:09.021526749 +0000 UTC m=+203.880039677" watchObservedRunningTime="2025-12-04 03:58:09.026528714 +0000 UTC m=+203.885041642" Dec 04 03:58:11 crc kubenswrapper[4806]: I1204 03:58:11.156494 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:58:11 crc kubenswrapper[4806]: I1204 03:58:11.157097 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:58:11 crc kubenswrapper[4806]: I1204 03:58:11.198912 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:58:11 crc kubenswrapper[4806]: I1204 03:58:11.555971 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:58:11 crc kubenswrapper[4806]: I1204 03:58:11.556053 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:58:11 crc kubenswrapper[4806]: I1204 03:58:11.598152 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:58:11 crc kubenswrapper[4806]: I1204 03:58:11.932602 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:58:11 crc kubenswrapper[4806]: I1204 03:58:11.932952 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:58:11 crc kubenswrapper[4806]: I1204 03:58:11.977595 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:58:12 crc kubenswrapper[4806]: I1204 03:58:12.056681 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:58:12 crc kubenswrapper[4806]: I1204 03:58:12.065781 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:58:12 crc kubenswrapper[4806]: I1204 03:58:12.065844 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:58:12 crc kubenswrapper[4806]: I1204 03:58:12.665413 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-brh2b"] Dec 04 03:58:13 crc kubenswrapper[4806]: I1204 03:58:13.309411 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:58:13 crc kubenswrapper[4806]: I1204 03:58:13.309462 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:58:13 crc kubenswrapper[4806]: I1204 03:58:13.347173 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:58:13 crc kubenswrapper[4806]: I1204 03:58:13.736780 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:58:13 crc kubenswrapper[4806]: I1204 03:58:13.736843 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:58:13 crc kubenswrapper[4806]: I1204 03:58:13.776236 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:58:14 crc kubenswrapper[4806]: I1204 03:58:14.009605 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-brh2b" podUID="07d1500e-9acd-47da-b1cb-13a7c7123f2a" containerName="registry-server" containerID="cri-o://9d7c7b1f48626d44501f40f7e834749b654026aaf5a15b1c2fa5807facfb542c" gracePeriod=2 Dec 04 03:58:14 crc kubenswrapper[4806]: I1204 03:58:14.056879 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:58:14 crc kubenswrapper[4806]: I1204 03:58:14.058667 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:58:14 crc kubenswrapper[4806]: I1204 03:58:14.345779 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:58:14 crc kubenswrapper[4806]: I1204 03:58:14.389642 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:58:14 crc kubenswrapper[4806]: I1204 03:58:14.758386 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:58:14 crc kubenswrapper[4806]: I1204 03:58:14.815508 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:58:14 crc kubenswrapper[4806]: I1204 03:58:14.896668 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:58:14 crc kubenswrapper[4806]: I1204 03:58:14.932989 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d1500e-9acd-47da-b1cb-13a7c7123f2a-catalog-content\") pod \"07d1500e-9acd-47da-b1cb-13a7c7123f2a\" (UID: \"07d1500e-9acd-47da-b1cb-13a7c7123f2a\") " Dec 04 03:58:14 crc kubenswrapper[4806]: I1204 03:58:14.933251 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d1500e-9acd-47da-b1cb-13a7c7123f2a-utilities\") pod \"07d1500e-9acd-47da-b1cb-13a7c7123f2a\" (UID: \"07d1500e-9acd-47da-b1cb-13a7c7123f2a\") " Dec 04 03:58:14 crc kubenswrapper[4806]: I1204 03:58:14.933336 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnddt\" (UniqueName: \"kubernetes.io/projected/07d1500e-9acd-47da-b1cb-13a7c7123f2a-kube-api-access-dnddt\") pod \"07d1500e-9acd-47da-b1cb-13a7c7123f2a\" (UID: \"07d1500e-9acd-47da-b1cb-13a7c7123f2a\") " Dec 04 03:58:14 crc kubenswrapper[4806]: I1204 03:58:14.937001 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07d1500e-9acd-47da-b1cb-13a7c7123f2a-utilities" (OuterVolumeSpecName: "utilities") pod "07d1500e-9acd-47da-b1cb-13a7c7123f2a" (UID: "07d1500e-9acd-47da-b1cb-13a7c7123f2a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:58:14 crc kubenswrapper[4806]: I1204 03:58:14.954275 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07d1500e-9acd-47da-b1cb-13a7c7123f2a-kube-api-access-dnddt" (OuterVolumeSpecName: "kube-api-access-dnddt") pod "07d1500e-9acd-47da-b1cb-13a7c7123f2a" (UID: "07d1500e-9acd-47da-b1cb-13a7c7123f2a"). InnerVolumeSpecName "kube-api-access-dnddt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:58:14 crc kubenswrapper[4806]: I1204 03:58:14.988585 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07d1500e-9acd-47da-b1cb-13a7c7123f2a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07d1500e-9acd-47da-b1cb-13a7c7123f2a" (UID: "07d1500e-9acd-47da-b1cb-13a7c7123f2a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.016589 4806 generic.go:334] "Generic (PLEG): container finished" podID="07d1500e-9acd-47da-b1cb-13a7c7123f2a" containerID="9d7c7b1f48626d44501f40f7e834749b654026aaf5a15b1c2fa5807facfb542c" exitCode=0 Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.016691 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brh2b" event={"ID":"07d1500e-9acd-47da-b1cb-13a7c7123f2a","Type":"ContainerDied","Data":"9d7c7b1f48626d44501f40f7e834749b654026aaf5a15b1c2fa5807facfb542c"} Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.016813 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-brh2b" Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.017131 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brh2b" event={"ID":"07d1500e-9acd-47da-b1cb-13a7c7123f2a","Type":"ContainerDied","Data":"e9a0398b52fa1880967552bfc2050383ed3cfed471193fa2462d9a1372f4ba9c"} Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.017187 4806 scope.go:117] "RemoveContainer" containerID="9d7c7b1f48626d44501f40f7e834749b654026aaf5a15b1c2fa5807facfb542c" Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.035583 4806 scope.go:117] "RemoveContainer" containerID="7a63f9ee49f8b4b935ce20874d60f000009fd9bb7ac7c65a5e2285c70b6412c2" Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.036932 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d1500e-9acd-47da-b1cb-13a7c7123f2a-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.037018 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnddt\" (UniqueName: \"kubernetes.io/projected/07d1500e-9acd-47da-b1cb-13a7c7123f2a-kube-api-access-dnddt\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.037035 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d1500e-9acd-47da-b1cb-13a7c7123f2a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.050272 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-brh2b"] Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.055361 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-brh2b"] Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.069818 4806 scope.go:117] "RemoveContainer" containerID="8de6a4d31028447c381ab0ba5d16e8b653b96a190f095f9886548410ba66738f" Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.085599 4806 scope.go:117] "RemoveContainer" containerID="9d7c7b1f48626d44501f40f7e834749b654026aaf5a15b1c2fa5807facfb542c" Dec 04 03:58:15 crc kubenswrapper[4806]: E1204 03:58:15.086339 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d7c7b1f48626d44501f40f7e834749b654026aaf5a15b1c2fa5807facfb542c\": container with ID starting with 9d7c7b1f48626d44501f40f7e834749b654026aaf5a15b1c2fa5807facfb542c not found: ID does not exist" containerID="9d7c7b1f48626d44501f40f7e834749b654026aaf5a15b1c2fa5807facfb542c" Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.086483 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d7c7b1f48626d44501f40f7e834749b654026aaf5a15b1c2fa5807facfb542c"} err="failed to get container status \"9d7c7b1f48626d44501f40f7e834749b654026aaf5a15b1c2fa5807facfb542c\": rpc error: code = NotFound desc = could not find container \"9d7c7b1f48626d44501f40f7e834749b654026aaf5a15b1c2fa5807facfb542c\": container with ID starting with 9d7c7b1f48626d44501f40f7e834749b654026aaf5a15b1c2fa5807facfb542c not found: ID does not exist" Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.086625 4806 scope.go:117] "RemoveContainer" containerID="7a63f9ee49f8b4b935ce20874d60f000009fd9bb7ac7c65a5e2285c70b6412c2" Dec 04 03:58:15 crc kubenswrapper[4806]: E1204 03:58:15.087352 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a63f9ee49f8b4b935ce20874d60f000009fd9bb7ac7c65a5e2285c70b6412c2\": container with ID starting with 7a63f9ee49f8b4b935ce20874d60f000009fd9bb7ac7c65a5e2285c70b6412c2 not found: ID does not exist" containerID="7a63f9ee49f8b4b935ce20874d60f000009fd9bb7ac7c65a5e2285c70b6412c2" Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.087424 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a63f9ee49f8b4b935ce20874d60f000009fd9bb7ac7c65a5e2285c70b6412c2"} err="failed to get container status \"7a63f9ee49f8b4b935ce20874d60f000009fd9bb7ac7c65a5e2285c70b6412c2\": rpc error: code = NotFound desc = could not find container \"7a63f9ee49f8b4b935ce20874d60f000009fd9bb7ac7c65a5e2285c70b6412c2\": container with ID starting with 7a63f9ee49f8b4b935ce20874d60f000009fd9bb7ac7c65a5e2285c70b6412c2 not found: ID does not exist" Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.087461 4806 scope.go:117] "RemoveContainer" containerID="8de6a4d31028447c381ab0ba5d16e8b653b96a190f095f9886548410ba66738f" Dec 04 03:58:15 crc kubenswrapper[4806]: E1204 03:58:15.087907 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8de6a4d31028447c381ab0ba5d16e8b653b96a190f095f9886548410ba66738f\": container with ID starting with 8de6a4d31028447c381ab0ba5d16e8b653b96a190f095f9886548410ba66738f not found: ID does not exist" containerID="8de6a4d31028447c381ab0ba5d16e8b653b96a190f095f9886548410ba66738f" Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.087976 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8de6a4d31028447c381ab0ba5d16e8b653b96a190f095f9886548410ba66738f"} err="failed to get container status \"8de6a4d31028447c381ab0ba5d16e8b653b96a190f095f9886548410ba66738f\": rpc error: code = NotFound desc = could not find container \"8de6a4d31028447c381ab0ba5d16e8b653b96a190f095f9886548410ba66738f\": container with ID starting with 8de6a4d31028447c381ab0ba5d16e8b653b96a190f095f9886548410ba66738f not found: ID does not exist" Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.432548 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07d1500e-9acd-47da-b1cb-13a7c7123f2a" path="/var/lib/kubelet/pods/07d1500e-9acd-47da-b1cb-13a7c7123f2a/volumes" Dec 04 03:58:15 crc kubenswrapper[4806]: I1204 03:58:15.961654 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" podUID="000c6858-a809-46e4-bdcb-b150477e5ce2" containerName="oauth-openshift" containerID="cri-o://2186f3a487247eca0a0153acc260f6eb90fd441fbe8e0427fc379a32eaa09295" gracePeriod=15 Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.304180 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.358437 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-serving-cert\") pod \"000c6858-a809-46e4-bdcb-b150477e5ce2\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.358497 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-trusted-ca-bundle\") pod \"000c6858-a809-46e4-bdcb-b150477e5ce2\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.358518 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-error\") pod \"000c6858-a809-46e4-bdcb-b150477e5ce2\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.358536 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-router-certs\") pod \"000c6858-a809-46e4-bdcb-b150477e5ce2\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.358557 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/000c6858-a809-46e4-bdcb-b150477e5ce2-audit-dir\") pod \"000c6858-a809-46e4-bdcb-b150477e5ce2\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.358574 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-provider-selection\") pod \"000c6858-a809-46e4-bdcb-b150477e5ce2\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.358619 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-ocp-branding-template\") pod \"000c6858-a809-46e4-bdcb-b150477e5ce2\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.358650 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-idp-0-file-data\") pod \"000c6858-a809-46e4-bdcb-b150477e5ce2\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.358666 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-cliconfig\") pod \"000c6858-a809-46e4-bdcb-b150477e5ce2\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.358673 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/000c6858-a809-46e4-bdcb-b150477e5ce2-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "000c6858-a809-46e4-bdcb-b150477e5ce2" (UID: "000c6858-a809-46e4-bdcb-b150477e5ce2"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.358700 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-service-ca\") pod \"000c6858-a809-46e4-bdcb-b150477e5ce2\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.358778 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6c5j\" (UniqueName: \"kubernetes.io/projected/000c6858-a809-46e4-bdcb-b150477e5ce2-kube-api-access-f6c5j\") pod \"000c6858-a809-46e4-bdcb-b150477e5ce2\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.358813 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-login\") pod \"000c6858-a809-46e4-bdcb-b150477e5ce2\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.358866 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-session\") pod \"000c6858-a809-46e4-bdcb-b150477e5ce2\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.358940 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-audit-policies\") pod \"000c6858-a809-46e4-bdcb-b150477e5ce2\" (UID: \"000c6858-a809-46e4-bdcb-b150477e5ce2\") " Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.359303 4806 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/000c6858-a809-46e4-bdcb-b150477e5ce2-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.359490 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "000c6858-a809-46e4-bdcb-b150477e5ce2" (UID: "000c6858-a809-46e4-bdcb-b150477e5ce2"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.359636 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "000c6858-a809-46e4-bdcb-b150477e5ce2" (UID: "000c6858-a809-46e4-bdcb-b150477e5ce2"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.359754 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "000c6858-a809-46e4-bdcb-b150477e5ce2" (UID: "000c6858-a809-46e4-bdcb-b150477e5ce2"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.359877 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "000c6858-a809-46e4-bdcb-b150477e5ce2" (UID: "000c6858-a809-46e4-bdcb-b150477e5ce2"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.364170 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "000c6858-a809-46e4-bdcb-b150477e5ce2" (UID: "000c6858-a809-46e4-bdcb-b150477e5ce2"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.364489 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "000c6858-a809-46e4-bdcb-b150477e5ce2" (UID: "000c6858-a809-46e4-bdcb-b150477e5ce2"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.364873 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "000c6858-a809-46e4-bdcb-b150477e5ce2" (UID: "000c6858-a809-46e4-bdcb-b150477e5ce2"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.365311 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "000c6858-a809-46e4-bdcb-b150477e5ce2" (UID: "000c6858-a809-46e4-bdcb-b150477e5ce2"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.365671 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "000c6858-a809-46e4-bdcb-b150477e5ce2" (UID: "000c6858-a809-46e4-bdcb-b150477e5ce2"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.366393 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "000c6858-a809-46e4-bdcb-b150477e5ce2" (UID: "000c6858-a809-46e4-bdcb-b150477e5ce2"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.367046 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "000c6858-a809-46e4-bdcb-b150477e5ce2" (UID: "000c6858-a809-46e4-bdcb-b150477e5ce2"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.367372 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "000c6858-a809-46e4-bdcb-b150477e5ce2" (UID: "000c6858-a809-46e4-bdcb-b150477e5ce2"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.372333 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/000c6858-a809-46e4-bdcb-b150477e5ce2-kube-api-access-f6c5j" (OuterVolumeSpecName: "kube-api-access-f6c5j") pod "000c6858-a809-46e4-bdcb-b150477e5ce2" (UID: "000c6858-a809-46e4-bdcb-b150477e5ce2"). InnerVolumeSpecName "kube-api-access-f6c5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.460800 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.460837 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.460854 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.460864 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6c5j\" (UniqueName: \"kubernetes.io/projected/000c6858-a809-46e4-bdcb-b150477e5ce2-kube-api-access-f6c5j\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.460874 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.460886 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.460899 4806 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.460910 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.460963 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.460974 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.460984 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.460997 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:16 crc kubenswrapper[4806]: I1204 03:58:16.461009 4806 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/000c6858-a809-46e4-bdcb-b150477e5ce2-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.032997 4806 generic.go:334] "Generic (PLEG): container finished" podID="000c6858-a809-46e4-bdcb-b150477e5ce2" containerID="2186f3a487247eca0a0153acc260f6eb90fd441fbe8e0427fc379a32eaa09295" exitCode=0 Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.033042 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" event={"ID":"000c6858-a809-46e4-bdcb-b150477e5ce2","Type":"ContainerDied","Data":"2186f3a487247eca0a0153acc260f6eb90fd441fbe8e0427fc379a32eaa09295"} Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.033067 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" event={"ID":"000c6858-a809-46e4-bdcb-b150477e5ce2","Type":"ContainerDied","Data":"c02c42375f4dea63a29a4d64a761689fc676c47d9015e254a8afe60f84919977"} Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.033084 4806 scope.go:117] "RemoveContainer" containerID="2186f3a487247eca0a0153acc260f6eb90fd441fbe8e0427fc379a32eaa09295" Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.033166 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7m42k" Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.062536 4806 scope.go:117] "RemoveContainer" containerID="2186f3a487247eca0a0153acc260f6eb90fd441fbe8e0427fc379a32eaa09295" Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.062850 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7m42k"] Dec 04 03:58:17 crc kubenswrapper[4806]: E1204 03:58:17.063023 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2186f3a487247eca0a0153acc260f6eb90fd441fbe8e0427fc379a32eaa09295\": container with ID starting with 2186f3a487247eca0a0153acc260f6eb90fd441fbe8e0427fc379a32eaa09295 not found: ID does not exist" containerID="2186f3a487247eca0a0153acc260f6eb90fd441fbe8e0427fc379a32eaa09295" Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.063053 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2186f3a487247eca0a0153acc260f6eb90fd441fbe8e0427fc379a32eaa09295"} err="failed to get container status \"2186f3a487247eca0a0153acc260f6eb90fd441fbe8e0427fc379a32eaa09295\": rpc error: code = NotFound desc = could not find container \"2186f3a487247eca0a0153acc260f6eb90fd441fbe8e0427fc379a32eaa09295\": container with ID starting with 2186f3a487247eca0a0153acc260f6eb90fd441fbe8e0427fc379a32eaa09295 not found: ID does not exist" Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.072822 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7m42k"] Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.431097 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="000c6858-a809-46e4-bdcb-b150477e5ce2" path="/var/lib/kubelet/pods/000c6858-a809-46e4-bdcb-b150477e5ce2/volumes" Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.462258 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjp4q"] Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.462512 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zjp4q" podUID="8d1951d0-8637-4110-ba57-ac3ee5f6c378" containerName="registry-server" containerID="cri-o://362f8314a2ffc4b44fbee193dad7a4161eaf444a32d6572e1d6664e60581e82e" gracePeriod=2 Dec 04 03:58:17 crc kubenswrapper[4806]: E1204 03:58:17.558302 4806 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d1951d0_8637_4110_ba57_ac3ee5f6c378.slice/crio-conmon-362f8314a2ffc4b44fbee193dad7a4161eaf444a32d6572e1d6664e60581e82e.scope\": RecentStats: unable to find data in memory cache]" Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.663072 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-74cvd"] Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.663330 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-74cvd" podUID="e792b62d-a7ed-4c50-8d30-90bb6b11b367" containerName="registry-server" containerID="cri-o://b5346e8b5afe8cd994f2b340e662d3c89ea5b253b2fec73f0c47f8c2673da6f3" gracePeriod=2 Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.817393 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.878024 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1951d0-8637-4110-ba57-ac3ee5f6c378-catalog-content\") pod \"8d1951d0-8637-4110-ba57-ac3ee5f6c378\" (UID: \"8d1951d0-8637-4110-ba57-ac3ee5f6c378\") " Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.878148 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1951d0-8637-4110-ba57-ac3ee5f6c378-utilities\") pod \"8d1951d0-8637-4110-ba57-ac3ee5f6c378\" (UID: \"8d1951d0-8637-4110-ba57-ac3ee5f6c378\") " Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.878176 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4574\" (UniqueName: \"kubernetes.io/projected/8d1951d0-8637-4110-ba57-ac3ee5f6c378-kube-api-access-h4574\") pod \"8d1951d0-8637-4110-ba57-ac3ee5f6c378\" (UID: \"8d1951d0-8637-4110-ba57-ac3ee5f6c378\") " Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.878921 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d1951d0-8637-4110-ba57-ac3ee5f6c378-utilities" (OuterVolumeSpecName: "utilities") pod "8d1951d0-8637-4110-ba57-ac3ee5f6c378" (UID: "8d1951d0-8637-4110-ba57-ac3ee5f6c378"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.884619 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d1951d0-8637-4110-ba57-ac3ee5f6c378-kube-api-access-h4574" (OuterVolumeSpecName: "kube-api-access-h4574") pod "8d1951d0-8637-4110-ba57-ac3ee5f6c378" (UID: "8d1951d0-8637-4110-ba57-ac3ee5f6c378"). InnerVolumeSpecName "kube-api-access-h4574". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.897443 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d1951d0-8637-4110-ba57-ac3ee5f6c378-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d1951d0-8637-4110-ba57-ac3ee5f6c378" (UID: "8d1951d0-8637-4110-ba57-ac3ee5f6c378"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.980040 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1951d0-8637-4110-ba57-ac3ee5f6c378-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.980089 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1951d0-8637-4110-ba57-ac3ee5f6c378-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:17 crc kubenswrapper[4806]: I1204 03:58:17.980145 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4574\" (UniqueName: \"kubernetes.io/projected/8d1951d0-8637-4110-ba57-ac3ee5f6c378-kube-api-access-h4574\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.039171 4806 generic.go:334] "Generic (PLEG): container finished" podID="8d1951d0-8637-4110-ba57-ac3ee5f6c378" containerID="362f8314a2ffc4b44fbee193dad7a4161eaf444a32d6572e1d6664e60581e82e" exitCode=0 Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.039579 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zjp4q" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.040735 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjp4q" event={"ID":"8d1951d0-8637-4110-ba57-ac3ee5f6c378","Type":"ContainerDied","Data":"362f8314a2ffc4b44fbee193dad7a4161eaf444a32d6572e1d6664e60581e82e"} Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.040808 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zjp4q" event={"ID":"8d1951d0-8637-4110-ba57-ac3ee5f6c378","Type":"ContainerDied","Data":"cf3a45ee2a81d2c903d4a27b28deebe8deacba41da9e4a81e4f1a75abfb8a466"} Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.040827 4806 scope.go:117] "RemoveContainer" containerID="362f8314a2ffc4b44fbee193dad7a4161eaf444a32d6572e1d6664e60581e82e" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.044236 4806 generic.go:334] "Generic (PLEG): container finished" podID="e792b62d-a7ed-4c50-8d30-90bb6b11b367" containerID="b5346e8b5afe8cd994f2b340e662d3c89ea5b253b2fec73f0c47f8c2673da6f3" exitCode=0 Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.044310 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74cvd" event={"ID":"e792b62d-a7ed-4c50-8d30-90bb6b11b367","Type":"ContainerDied","Data":"b5346e8b5afe8cd994f2b340e662d3c89ea5b253b2fec73f0c47f8c2673da6f3"} Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.071526 4806 scope.go:117] "RemoveContainer" containerID="1067ca88719c50540d0bacf11167c7c884852a6f3c739bd3334efb714f20d3f9" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.072775 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjp4q"] Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.074726 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zjp4q"] Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.091999 4806 scope.go:117] "RemoveContainer" containerID="b88f4c91a83516dd76144063318ec60ea57a4f5ced9762e29a6af263a0c67cba" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.095044 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.107899 4806 scope.go:117] "RemoveContainer" containerID="362f8314a2ffc4b44fbee193dad7a4161eaf444a32d6572e1d6664e60581e82e" Dec 04 03:58:18 crc kubenswrapper[4806]: E1204 03:58:18.108397 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"362f8314a2ffc4b44fbee193dad7a4161eaf444a32d6572e1d6664e60581e82e\": container with ID starting with 362f8314a2ffc4b44fbee193dad7a4161eaf444a32d6572e1d6664e60581e82e not found: ID does not exist" containerID="362f8314a2ffc4b44fbee193dad7a4161eaf444a32d6572e1d6664e60581e82e" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.108435 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"362f8314a2ffc4b44fbee193dad7a4161eaf444a32d6572e1d6664e60581e82e"} err="failed to get container status \"362f8314a2ffc4b44fbee193dad7a4161eaf444a32d6572e1d6664e60581e82e\": rpc error: code = NotFound desc = could not find container \"362f8314a2ffc4b44fbee193dad7a4161eaf444a32d6572e1d6664e60581e82e\": container with ID starting with 362f8314a2ffc4b44fbee193dad7a4161eaf444a32d6572e1d6664e60581e82e not found: ID does not exist" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.108473 4806 scope.go:117] "RemoveContainer" containerID="1067ca88719c50540d0bacf11167c7c884852a6f3c739bd3334efb714f20d3f9" Dec 04 03:58:18 crc kubenswrapper[4806]: E1204 03:58:18.108789 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1067ca88719c50540d0bacf11167c7c884852a6f3c739bd3334efb714f20d3f9\": container with ID starting with 1067ca88719c50540d0bacf11167c7c884852a6f3c739bd3334efb714f20d3f9 not found: ID does not exist" containerID="1067ca88719c50540d0bacf11167c7c884852a6f3c739bd3334efb714f20d3f9" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.108830 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1067ca88719c50540d0bacf11167c7c884852a6f3c739bd3334efb714f20d3f9"} err="failed to get container status \"1067ca88719c50540d0bacf11167c7c884852a6f3c739bd3334efb714f20d3f9\": rpc error: code = NotFound desc = could not find container \"1067ca88719c50540d0bacf11167c7c884852a6f3c739bd3334efb714f20d3f9\": container with ID starting with 1067ca88719c50540d0bacf11167c7c884852a6f3c739bd3334efb714f20d3f9 not found: ID does not exist" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.108856 4806 scope.go:117] "RemoveContainer" containerID="b88f4c91a83516dd76144063318ec60ea57a4f5ced9762e29a6af263a0c67cba" Dec 04 03:58:18 crc kubenswrapper[4806]: E1204 03:58:18.109120 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b88f4c91a83516dd76144063318ec60ea57a4f5ced9762e29a6af263a0c67cba\": container with ID starting with b88f4c91a83516dd76144063318ec60ea57a4f5ced9762e29a6af263a0c67cba not found: ID does not exist" containerID="b88f4c91a83516dd76144063318ec60ea57a4f5ced9762e29a6af263a0c67cba" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.109144 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b88f4c91a83516dd76144063318ec60ea57a4f5ced9762e29a6af263a0c67cba"} err="failed to get container status \"b88f4c91a83516dd76144063318ec60ea57a4f5ced9762e29a6af263a0c67cba\": rpc error: code = NotFound desc = could not find container \"b88f4c91a83516dd76144063318ec60ea57a4f5ced9762e29a6af263a0c67cba\": container with ID starting with b88f4c91a83516dd76144063318ec60ea57a4f5ced9762e29a6af263a0c67cba not found: ID does not exist" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.181848 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brkj8\" (UniqueName: \"kubernetes.io/projected/e792b62d-a7ed-4c50-8d30-90bb6b11b367-kube-api-access-brkj8\") pod \"e792b62d-a7ed-4c50-8d30-90bb6b11b367\" (UID: \"e792b62d-a7ed-4c50-8d30-90bb6b11b367\") " Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.181903 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e792b62d-a7ed-4c50-8d30-90bb6b11b367-utilities\") pod \"e792b62d-a7ed-4c50-8d30-90bb6b11b367\" (UID: \"e792b62d-a7ed-4c50-8d30-90bb6b11b367\") " Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.181994 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e792b62d-a7ed-4c50-8d30-90bb6b11b367-catalog-content\") pod \"e792b62d-a7ed-4c50-8d30-90bb6b11b367\" (UID: \"e792b62d-a7ed-4c50-8d30-90bb6b11b367\") " Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.182793 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e792b62d-a7ed-4c50-8d30-90bb6b11b367-utilities" (OuterVolumeSpecName: "utilities") pod "e792b62d-a7ed-4c50-8d30-90bb6b11b367" (UID: "e792b62d-a7ed-4c50-8d30-90bb6b11b367"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.186334 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e792b62d-a7ed-4c50-8d30-90bb6b11b367-kube-api-access-brkj8" (OuterVolumeSpecName: "kube-api-access-brkj8") pod "e792b62d-a7ed-4c50-8d30-90bb6b11b367" (UID: "e792b62d-a7ed-4c50-8d30-90bb6b11b367"). InnerVolumeSpecName "kube-api-access-brkj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.281732 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e792b62d-a7ed-4c50-8d30-90bb6b11b367-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e792b62d-a7ed-4c50-8d30-90bb6b11b367" (UID: "e792b62d-a7ed-4c50-8d30-90bb6b11b367"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.282555 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brkj8\" (UniqueName: \"kubernetes.io/projected/e792b62d-a7ed-4c50-8d30-90bb6b11b367-kube-api-access-brkj8\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.282809 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e792b62d-a7ed-4c50-8d30-90bb6b11b367-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:18 crc kubenswrapper[4806]: I1204 03:58:18.282902 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e792b62d-a7ed-4c50-8d30-90bb6b11b367-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:19 crc kubenswrapper[4806]: I1204 03:58:19.052720 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74cvd" event={"ID":"e792b62d-a7ed-4c50-8d30-90bb6b11b367","Type":"ContainerDied","Data":"b5d87e0a1aacdf3b70102f65783d09327280d6ed70c0416e59a1589ae463dda5"} Dec 04 03:58:19 crc kubenswrapper[4806]: I1204 03:58:19.052774 4806 scope.go:117] "RemoveContainer" containerID="b5346e8b5afe8cd994f2b340e662d3c89ea5b253b2fec73f0c47f8c2673da6f3" Dec 04 03:58:19 crc kubenswrapper[4806]: I1204 03:58:19.052891 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74cvd" Dec 04 03:58:19 crc kubenswrapper[4806]: I1204 03:58:19.069048 4806 scope.go:117] "RemoveContainer" containerID="47927e658b525cfd6f895efb0a00d03af4078239d7c763e31c023a7a41a6dbe3" Dec 04 03:58:19 crc kubenswrapper[4806]: I1204 03:58:19.088117 4806 scope.go:117] "RemoveContainer" containerID="27c810f908ab717540af22e500b0604e032f6208bafceb77e0bdf1eb2e052fa3" Dec 04 03:58:19 crc kubenswrapper[4806]: I1204 03:58:19.092369 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-74cvd"] Dec 04 03:58:19 crc kubenswrapper[4806]: I1204 03:58:19.096208 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-74cvd"] Dec 04 03:58:19 crc kubenswrapper[4806]: I1204 03:58:19.433752 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d1951d0-8637-4110-ba57-ac3ee5f6c378" path="/var/lib/kubelet/pods/8d1951d0-8637-4110-ba57-ac3ee5f6c378/volumes" Dec 04 03:58:19 crc kubenswrapper[4806]: I1204 03:58:19.434976 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e792b62d-a7ed-4c50-8d30-90bb6b11b367" path="/var/lib/kubelet/pods/e792b62d-a7ed-4c50-8d30-90bb6b11b367/volumes" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.431851 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5549b7d775-j2p46"] Dec 04 03:58:23 crc kubenswrapper[4806]: E1204 03:58:23.432665 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85470b07-1f66-4a8a-b30c-85c2728ab076" containerName="extract-utilities" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.432685 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="85470b07-1f66-4a8a-b30c-85c2728ab076" containerName="extract-utilities" Dec 04 03:58:23 crc kubenswrapper[4806]: E1204 03:58:23.432696 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1951d0-8637-4110-ba57-ac3ee5f6c378" containerName="extract-content" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.432704 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1951d0-8637-4110-ba57-ac3ee5f6c378" containerName="extract-content" Dec 04 03:58:23 crc kubenswrapper[4806]: E1204 03:58:23.432721 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e792b62d-a7ed-4c50-8d30-90bb6b11b367" containerName="extract-utilities" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.432730 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e792b62d-a7ed-4c50-8d30-90bb6b11b367" containerName="extract-utilities" Dec 04 03:58:23 crc kubenswrapper[4806]: E1204 03:58:23.432743 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e792b62d-a7ed-4c50-8d30-90bb6b11b367" containerName="extract-content" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.432752 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e792b62d-a7ed-4c50-8d30-90bb6b11b367" containerName="extract-content" Dec 04 03:58:23 crc kubenswrapper[4806]: E1204 03:58:23.432762 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1951d0-8637-4110-ba57-ac3ee5f6c378" containerName="registry-server" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.432771 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1951d0-8637-4110-ba57-ac3ee5f6c378" containerName="registry-server" Dec 04 03:58:23 crc kubenswrapper[4806]: E1204 03:58:23.432781 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d1500e-9acd-47da-b1cb-13a7c7123f2a" containerName="extract-content" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.432789 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d1500e-9acd-47da-b1cb-13a7c7123f2a" containerName="extract-content" Dec 04 03:58:23 crc kubenswrapper[4806]: E1204 03:58:23.432803 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85470b07-1f66-4a8a-b30c-85c2728ab076" containerName="extract-content" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.432813 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="85470b07-1f66-4a8a-b30c-85c2728ab076" containerName="extract-content" Dec 04 03:58:23 crc kubenswrapper[4806]: E1204 03:58:23.432821 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d1951d0-8637-4110-ba57-ac3ee5f6c378" containerName="extract-utilities" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.432850 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d1951d0-8637-4110-ba57-ac3ee5f6c378" containerName="extract-utilities" Dec 04 03:58:23 crc kubenswrapper[4806]: E1204 03:58:23.432864 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000c6858-a809-46e4-bdcb-b150477e5ce2" containerName="oauth-openshift" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.432872 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="000c6858-a809-46e4-bdcb-b150477e5ce2" containerName="oauth-openshift" Dec 04 03:58:23 crc kubenswrapper[4806]: E1204 03:58:23.432883 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d1500e-9acd-47da-b1cb-13a7c7123f2a" containerName="registry-server" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.432890 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d1500e-9acd-47da-b1cb-13a7c7123f2a" containerName="registry-server" Dec 04 03:58:23 crc kubenswrapper[4806]: E1204 03:58:23.432900 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85470b07-1f66-4a8a-b30c-85c2728ab076" containerName="registry-server" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.432908 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="85470b07-1f66-4a8a-b30c-85c2728ab076" containerName="registry-server" Dec 04 03:58:23 crc kubenswrapper[4806]: E1204 03:58:23.432939 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e792b62d-a7ed-4c50-8d30-90bb6b11b367" containerName="registry-server" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.432950 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e792b62d-a7ed-4c50-8d30-90bb6b11b367" containerName="registry-server" Dec 04 03:58:23 crc kubenswrapper[4806]: E1204 03:58:23.432960 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d1500e-9acd-47da-b1cb-13a7c7123f2a" containerName="extract-utilities" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.432968 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d1500e-9acd-47da-b1cb-13a7c7123f2a" containerName="extract-utilities" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.433137 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="07d1500e-9acd-47da-b1cb-13a7c7123f2a" containerName="registry-server" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.433149 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d1951d0-8637-4110-ba57-ac3ee5f6c378" containerName="registry-server" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.433170 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="85470b07-1f66-4a8a-b30c-85c2728ab076" containerName="registry-server" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.433182 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e792b62d-a7ed-4c50-8d30-90bb6b11b367" containerName="registry-server" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.433193 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="000c6858-a809-46e4-bdcb-b150477e5ce2" containerName="oauth-openshift" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.433903 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.437408 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.437528 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.437621 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.437938 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.440420 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.440448 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.441102 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.441203 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.442197 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.442462 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.442836 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.443782 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.444813 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.444978 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ddc8257d-4491-4d88-afc4-85b72af1fbff-audit-policies\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.445183 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.445301 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.445417 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-user-template-login\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.445521 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-service-ca\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.445668 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.445861 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl2ms\" (UniqueName: \"kubernetes.io/projected/ddc8257d-4491-4d88-afc4-85b72af1fbff-kube-api-access-jl2ms\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.446044 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ddc8257d-4491-4d88-afc4-85b72af1fbff-audit-dir\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.446167 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.446297 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-router-certs\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.446414 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-session\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.446544 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.446745 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-user-template-error\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.458175 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.461446 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.470740 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5549b7d775-j2p46"] Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.471654 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.548236 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.548321 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-user-template-login\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.548352 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-service-ca\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.548389 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.548462 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl2ms\" (UniqueName: \"kubernetes.io/projected/ddc8257d-4491-4d88-afc4-85b72af1fbff-kube-api-access-jl2ms\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.549248 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.549248 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-service-ca\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.549328 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ddc8257d-4491-4d88-afc4-85b72af1fbff-audit-dir\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.549372 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.549400 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-session\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.549423 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-router-certs\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.549469 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.549501 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-user-template-error\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.549540 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.549563 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ddc8257d-4491-4d88-afc4-85b72af1fbff-audit-policies\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.549594 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.558570 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.563204 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.574450 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-user-template-login\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.574646 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ddc8257d-4491-4d88-afc4-85b72af1fbff-audit-dir\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.593949 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ddc8257d-4491-4d88-afc4-85b72af1fbff-audit-policies\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.594505 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.594834 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-user-template-error\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.596501 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-session\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.597166 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.600801 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.613552 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ddc8257d-4491-4d88-afc4-85b72af1fbff-v4-0-config-system-router-certs\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.614167 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl2ms\" (UniqueName: \"kubernetes.io/projected/ddc8257d-4491-4d88-afc4-85b72af1fbff-kube-api-access-jl2ms\") pod \"oauth-openshift-5549b7d775-j2p46\" (UID: \"ddc8257d-4491-4d88-afc4-85b72af1fbff\") " pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:23 crc kubenswrapper[4806]: I1204 03:58:23.756799 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:24 crc kubenswrapper[4806]: I1204 03:58:24.053517 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5549b7d775-j2p46"] Dec 04 03:58:24 crc kubenswrapper[4806]: I1204 03:58:24.084024 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" event={"ID":"ddc8257d-4491-4d88-afc4-85b72af1fbff","Type":"ContainerStarted","Data":"2da5b1f2ba63549369af1769181a433d1bad1970c634032e40ed6c8ede82ef55"} Dec 04 03:58:25 crc kubenswrapper[4806]: I1204 03:58:25.092263 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" event={"ID":"ddc8257d-4491-4d88-afc4-85b72af1fbff","Type":"ContainerStarted","Data":"df895389df62f5f57d4b651453b8ed777b969e39e1125167ed9c7ac4344ba8c9"} Dec 04 03:58:25 crc kubenswrapper[4806]: I1204 03:58:25.093793 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:25 crc kubenswrapper[4806]: I1204 03:58:25.102941 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" Dec 04 03:58:25 crc kubenswrapper[4806]: I1204 03:58:25.159515 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5549b7d775-j2p46" podStartSLOduration=35.159492122 podStartE2EDuration="35.159492122s" podCreationTimestamp="2025-12-04 03:57:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:58:25.12920565 +0000 UTC m=+219.987718578" watchObservedRunningTime="2025-12-04 03:58:25.159492122 +0000 UTC m=+220.018005050" Dec 04 03:58:27 crc kubenswrapper[4806]: I1204 03:58:27.047622 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 03:58:27 crc kubenswrapper[4806]: I1204 03:58:27.048076 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 03:58:27 crc kubenswrapper[4806]: I1204 03:58:27.048136 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 03:58:27 crc kubenswrapper[4806]: I1204 03:58:27.048752 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 03:58:27 crc kubenswrapper[4806]: I1204 03:58:27.048814 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d" gracePeriod=600 Dec 04 03:58:28 crc kubenswrapper[4806]: I1204 03:58:28.109881 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d" exitCode=0 Dec 04 03:58:28 crc kubenswrapper[4806]: I1204 03:58:28.109963 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d"} Dec 04 03:58:28 crc kubenswrapper[4806]: I1204 03:58:28.110563 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"3e4ceeaa6204e4c1f2d29168fb61b4f1c4c0b7afb94d21facd0f0961ea7e32ff"} Dec 04 03:58:38 crc kubenswrapper[4806]: I1204 03:58:38.826796 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q9fqz"] Dec 04 03:58:38 crc kubenswrapper[4806]: I1204 03:58:38.831233 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q9fqz" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" containerName="registry-server" containerID="cri-o://04dd0126cf689d5f7aa3f5021261660f9566c1317f784ebbc0a0e5b90f9a74df" gracePeriod=30 Dec 04 03:58:38 crc kubenswrapper[4806]: I1204 03:58:38.838227 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rlrnj"] Dec 04 03:58:38 crc kubenswrapper[4806]: I1204 03:58:38.838497 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rlrnj" podUID="21c59133-1783-4225-a70d-1395d9642980" containerName="registry-server" containerID="cri-o://84659a3af5d80e162ecbe638130986ac93820569328f6cb1a5f7a30d3d198440" gracePeriod=30 Dec 04 03:58:38 crc kubenswrapper[4806]: I1204 03:58:38.847210 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mzgjp"] Dec 04 03:58:38 crc kubenswrapper[4806]: I1204 03:58:38.847480 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" podUID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" containerName="marketplace-operator" containerID="cri-o://93979e497259d4fb735040c8bb60e0dbbf4d26aa5dad45718f9071972f3929ac" gracePeriod=30 Dec 04 03:58:38 crc kubenswrapper[4806]: I1204 03:58:38.858848 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ln2r5"] Dec 04 03:58:38 crc kubenswrapper[4806]: I1204 03:58:38.859145 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ln2r5" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" containerName="registry-server" containerID="cri-o://3d8a89e3e4eb267dcc1cb982fa7ce09d4b6fe168fad8c21218c0de647968c8f9" gracePeriod=30 Dec 04 03:58:38 crc kubenswrapper[4806]: I1204 03:58:38.880691 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hwdhx"] Dec 04 03:58:38 crc kubenswrapper[4806]: I1204 03:58:38.880975 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hwdhx" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" containerName="registry-server" containerID="cri-o://aa4c2ceb9165d54155139b13b8e3dd58f12f8aa97ba2236faa3c276ca56eb542" gracePeriod=30 Dec 04 03:58:38 crc kubenswrapper[4806]: I1204 03:58:38.891029 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9rvth"] Dec 04 03:58:38 crc kubenswrapper[4806]: I1204 03:58:38.891815 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:38 crc kubenswrapper[4806]: I1204 03:58:38.904634 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9rvth"] Dec 04 03:58:38 crc kubenswrapper[4806]: I1204 03:58:38.976565 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/60ce8b7c-5cf6-452c-835e-9ab84a66aa65-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9rvth\" (UID: \"60ce8b7c-5cf6-452c-835e-9ab84a66aa65\") " pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:38 crc kubenswrapper[4806]: I1204 03:58:38.976629 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/60ce8b7c-5cf6-452c-835e-9ab84a66aa65-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9rvth\" (UID: \"60ce8b7c-5cf6-452c-835e-9ab84a66aa65\") " pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:38 crc kubenswrapper[4806]: I1204 03:58:38.976666 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h255q\" (UniqueName: \"kubernetes.io/projected/60ce8b7c-5cf6-452c-835e-9ab84a66aa65-kube-api-access-h255q\") pod \"marketplace-operator-79b997595-9rvth\" (UID: \"60ce8b7c-5cf6-452c-835e-9ab84a66aa65\") " pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.078005 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/60ce8b7c-5cf6-452c-835e-9ab84a66aa65-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9rvth\" (UID: \"60ce8b7c-5cf6-452c-835e-9ab84a66aa65\") " pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.078113 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h255q\" (UniqueName: \"kubernetes.io/projected/60ce8b7c-5cf6-452c-835e-9ab84a66aa65-kube-api-access-h255q\") pod \"marketplace-operator-79b997595-9rvth\" (UID: \"60ce8b7c-5cf6-452c-835e-9ab84a66aa65\") " pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.078897 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/60ce8b7c-5cf6-452c-835e-9ab84a66aa65-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9rvth\" (UID: \"60ce8b7c-5cf6-452c-835e-9ab84a66aa65\") " pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.082201 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/60ce8b7c-5cf6-452c-835e-9ab84a66aa65-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9rvth\" (UID: \"60ce8b7c-5cf6-452c-835e-9ab84a66aa65\") " pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.091819 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/60ce8b7c-5cf6-452c-835e-9ab84a66aa65-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9rvth\" (UID: \"60ce8b7c-5cf6-452c-835e-9ab84a66aa65\") " pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.101772 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h255q\" (UniqueName: \"kubernetes.io/projected/60ce8b7c-5cf6-452c-835e-9ab84a66aa65-kube-api-access-h255q\") pod \"marketplace-operator-79b997595-9rvth\" (UID: \"60ce8b7c-5cf6-452c-835e-9ab84a66aa65\") " pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.190314 4806 generic.go:334] "Generic (PLEG): container finished" podID="59a2e27f-ed83-42ba-9371-14546fbcc6af" containerID="04dd0126cf689d5f7aa3f5021261660f9566c1317f784ebbc0a0e5b90f9a74df" exitCode=0 Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.190374 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q9fqz" event={"ID":"59a2e27f-ed83-42ba-9371-14546fbcc6af","Type":"ContainerDied","Data":"04dd0126cf689d5f7aa3f5021261660f9566c1317f784ebbc0a0e5b90f9a74df"} Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.193364 4806 generic.go:334] "Generic (PLEG): container finished" podID="21c59133-1783-4225-a70d-1395d9642980" containerID="84659a3af5d80e162ecbe638130986ac93820569328f6cb1a5f7a30d3d198440" exitCode=0 Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.193400 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlrnj" event={"ID":"21c59133-1783-4225-a70d-1395d9642980","Type":"ContainerDied","Data":"84659a3af5d80e162ecbe638130986ac93820569328f6cb1a5f7a30d3d198440"} Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.195020 4806 generic.go:334] "Generic (PLEG): container finished" podID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" containerID="93979e497259d4fb735040c8bb60e0dbbf4d26aa5dad45718f9071972f3929ac" exitCode=0 Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.195051 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" event={"ID":"4e4767a8-ff51-43c7-94cc-625c9b8689a3","Type":"ContainerDied","Data":"93979e497259d4fb735040c8bb60e0dbbf4d26aa5dad45718f9071972f3929ac"} Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.196704 4806 generic.go:334] "Generic (PLEG): container finished" podID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" containerID="3d8a89e3e4eb267dcc1cb982fa7ce09d4b6fe168fad8c21218c0de647968c8f9" exitCode=0 Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.196745 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ln2r5" event={"ID":"ef7e1243-9a57-4bbc-a28a-43cb93cf2add","Type":"ContainerDied","Data":"3d8a89e3e4eb267dcc1cb982fa7ce09d4b6fe168fad8c21218c0de647968c8f9"} Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.202498 4806 generic.go:334] "Generic (PLEG): container finished" podID="eaca0644-a305-41b3-84f1-4179e3412d37" containerID="aa4c2ceb9165d54155139b13b8e3dd58f12f8aa97ba2236faa3c276ca56eb542" exitCode=0 Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.202536 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwdhx" event={"ID":"eaca0644-a305-41b3-84f1-4179e3412d37","Type":"ContainerDied","Data":"aa4c2ceb9165d54155139b13b8e3dd58f12f8aa97ba2236faa3c276ca56eb542"} Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.216254 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.306332 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.384943 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaca0644-a305-41b3-84f1-4179e3412d37-utilities\") pod \"eaca0644-a305-41b3-84f1-4179e3412d37\" (UID: \"eaca0644-a305-41b3-84f1-4179e3412d37\") " Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.385122 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94gnq\" (UniqueName: \"kubernetes.io/projected/eaca0644-a305-41b3-84f1-4179e3412d37-kube-api-access-94gnq\") pod \"eaca0644-a305-41b3-84f1-4179e3412d37\" (UID: \"eaca0644-a305-41b3-84f1-4179e3412d37\") " Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.385172 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaca0644-a305-41b3-84f1-4179e3412d37-catalog-content\") pod \"eaca0644-a305-41b3-84f1-4179e3412d37\" (UID: \"eaca0644-a305-41b3-84f1-4179e3412d37\") " Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.389997 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaca0644-a305-41b3-84f1-4179e3412d37-utilities" (OuterVolumeSpecName: "utilities") pod "eaca0644-a305-41b3-84f1-4179e3412d37" (UID: "eaca0644-a305-41b3-84f1-4179e3412d37"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.397889 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaca0644-a305-41b3-84f1-4179e3412d37-kube-api-access-94gnq" (OuterVolumeSpecName: "kube-api-access-94gnq") pod "eaca0644-a305-41b3-84f1-4179e3412d37" (UID: "eaca0644-a305-41b3-84f1-4179e3412d37"). InnerVolumeSpecName "kube-api-access-94gnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.468652 4806 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 03:58:39 crc kubenswrapper[4806]: E1204 03:58:39.468947 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" containerName="registry-server" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.468964 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" containerName="registry-server" Dec 04 03:58:39 crc kubenswrapper[4806]: E1204 03:58:39.468979 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" containerName="extract-utilities" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.468987 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" containerName="extract-utilities" Dec 04 03:58:39 crc kubenswrapper[4806]: E1204 03:58:39.469006 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" containerName="extract-content" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.469013 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" containerName="extract-content" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.469165 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" containerName="registry-server" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.469583 4806 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.469740 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.470303 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da" gracePeriod=15 Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.470543 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617" gracePeriod=15 Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.470570 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9" gracePeriod=15 Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.470859 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909" gracePeriod=15 Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.470630 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44" gracePeriod=15 Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.473779 4806 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 03:58:39 crc kubenswrapper[4806]: E1204 03:58:39.480736 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.480766 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 03:58:39 crc kubenswrapper[4806]: E1204 03:58:39.480776 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.480783 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 03:58:39 crc kubenswrapper[4806]: E1204 03:58:39.480792 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.480799 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 03:58:39 crc kubenswrapper[4806]: E1204 03:58:39.480810 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.480816 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 04 03:58:39 crc kubenswrapper[4806]: E1204 03:58:39.480825 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.480831 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 03:58:39 crc kubenswrapper[4806]: E1204 03:58:39.480845 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.480853 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.480977 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.480992 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.481006 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.481016 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.481026 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.486398 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaca0644-a305-41b3-84f1-4179e3412d37-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.486451 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94gnq\" (UniqueName: \"kubernetes.io/projected/eaca0644-a305-41b3-84f1-4179e3412d37-kube-api-access-94gnq\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.509282 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.587894 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5m5h\" (UniqueName: \"kubernetes.io/projected/4e4767a8-ff51-43c7-94cc-625c9b8689a3-kube-api-access-j5m5h\") pod \"4e4767a8-ff51-43c7-94cc-625c9b8689a3\" (UID: \"4e4767a8-ff51-43c7-94cc-625c9b8689a3\") " Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.588389 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4e4767a8-ff51-43c7-94cc-625c9b8689a3-marketplace-trusted-ca\") pod \"4e4767a8-ff51-43c7-94cc-625c9b8689a3\" (UID: \"4e4767a8-ff51-43c7-94cc-625c9b8689a3\") " Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.588425 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4e4767a8-ff51-43c7-94cc-625c9b8689a3-marketplace-operator-metrics\") pod \"4e4767a8-ff51-43c7-94cc-625c9b8689a3\" (UID: \"4e4767a8-ff51-43c7-94cc-625c9b8689a3\") " Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.588585 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.588634 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.588687 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.588722 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.588749 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.588773 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.588823 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.588845 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.591047 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e4767a8-ff51-43c7-94cc-625c9b8689a3-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "4e4767a8-ff51-43c7-94cc-625c9b8689a3" (UID: "4e4767a8-ff51-43c7-94cc-625c9b8689a3"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.597378 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e4767a8-ff51-43c7-94cc-625c9b8689a3-kube-api-access-j5m5h" (OuterVolumeSpecName: "kube-api-access-j5m5h") pod "4e4767a8-ff51-43c7-94cc-625c9b8689a3" (UID: "4e4767a8-ff51-43c7-94cc-625c9b8689a3"). InnerVolumeSpecName "kube-api-access-j5m5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.597667 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e4767a8-ff51-43c7-94cc-625c9b8689a3-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "4e4767a8-ff51-43c7-94cc-625c9b8689a3" (UID: "4e4767a8-ff51-43c7-94cc-625c9b8689a3"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.650399 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.650976 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690133 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqpw7\" (UniqueName: \"kubernetes.io/projected/59a2e27f-ed83-42ba-9371-14546fbcc6af-kube-api-access-pqpw7\") pod \"59a2e27f-ed83-42ba-9371-14546fbcc6af\" (UID: \"59a2e27f-ed83-42ba-9371-14546fbcc6af\") " Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690197 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59a2e27f-ed83-42ba-9371-14546fbcc6af-utilities\") pod \"59a2e27f-ed83-42ba-9371-14546fbcc6af\" (UID: \"59a2e27f-ed83-42ba-9371-14546fbcc6af\") " Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690243 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-catalog-content\") pod \"ef7e1243-9a57-4bbc-a28a-43cb93cf2add\" (UID: \"ef7e1243-9a57-4bbc-a28a-43cb93cf2add\") " Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690272 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59a2e27f-ed83-42ba-9371-14546fbcc6af-catalog-content\") pod \"59a2e27f-ed83-42ba-9371-14546fbcc6af\" (UID: \"59a2e27f-ed83-42ba-9371-14546fbcc6af\") " Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690301 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf6x7\" (UniqueName: \"kubernetes.io/projected/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-kube-api-access-wf6x7\") pod \"ef7e1243-9a57-4bbc-a28a-43cb93cf2add\" (UID: \"ef7e1243-9a57-4bbc-a28a-43cb93cf2add\") " Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690319 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-utilities\") pod \"ef7e1243-9a57-4bbc-a28a-43cb93cf2add\" (UID: \"ef7e1243-9a57-4bbc-a28a-43cb93cf2add\") " Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690434 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690464 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690486 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690508 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690550 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690565 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690582 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690600 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690646 4806 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4e4767a8-ff51-43c7-94cc-625c9b8689a3-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690657 4806 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4e4767a8-ff51-43c7-94cc-625c9b8689a3-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690666 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5m5h\" (UniqueName: \"kubernetes.io/projected/4e4767a8-ff51-43c7-94cc-625c9b8689a3-kube-api-access-j5m5h\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.690705 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.691169 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.691259 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.691341 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.691378 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.691812 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59a2e27f-ed83-42ba-9371-14546fbcc6af-utilities" (OuterVolumeSpecName: "utilities") pod "59a2e27f-ed83-42ba-9371-14546fbcc6af" (UID: "59a2e27f-ed83-42ba-9371-14546fbcc6af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.693973 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.694026 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.694053 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.694622 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-utilities" (OuterVolumeSpecName: "utilities") pod "ef7e1243-9a57-4bbc-a28a-43cb93cf2add" (UID: "ef7e1243-9a57-4bbc-a28a-43cb93cf2add"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.698679 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaca0644-a305-41b3-84f1-4179e3412d37-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eaca0644-a305-41b3-84f1-4179e3412d37" (UID: "eaca0644-a305-41b3-84f1-4179e3412d37"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.722555 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-kube-api-access-wf6x7" (OuterVolumeSpecName: "kube-api-access-wf6x7") pod "ef7e1243-9a57-4bbc-a28a-43cb93cf2add" (UID: "ef7e1243-9a57-4bbc-a28a-43cb93cf2add"). InnerVolumeSpecName "kube-api-access-wf6x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.722604 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59a2e27f-ed83-42ba-9371-14546fbcc6af-kube-api-access-pqpw7" (OuterVolumeSpecName: "kube-api-access-pqpw7") pod "59a2e27f-ed83-42ba-9371-14546fbcc6af" (UID: "59a2e27f-ed83-42ba-9371-14546fbcc6af"). InnerVolumeSpecName "kube-api-access-pqpw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.746395 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ef7e1243-9a57-4bbc-a28a-43cb93cf2add" (UID: "ef7e1243-9a57-4bbc-a28a-43cb93cf2add"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.757258 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59a2e27f-ed83-42ba-9371-14546fbcc6af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59a2e27f-ed83-42ba-9371-14546fbcc6af" (UID: "59a2e27f-ed83-42ba-9371-14546fbcc6af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.802278 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59a2e27f-ed83-42ba-9371-14546fbcc6af-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.802502 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaca0644-a305-41b3-84f1-4179e3412d37-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.802825 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.802870 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59a2e27f-ed83-42ba-9371-14546fbcc6af-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.802883 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.802894 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf6x7\" (UniqueName: \"kubernetes.io/projected/ef7e1243-9a57-4bbc-a28a-43cb93cf2add-kube-api-access-wf6x7\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.802906 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqpw7\" (UniqueName: \"kubernetes.io/projected/59a2e27f-ed83-42ba-9371-14546fbcc6af-kube-api-access-pqpw7\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.853559 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.903658 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21c59133-1783-4225-a70d-1395d9642980-utilities\") pod \"21c59133-1783-4225-a70d-1395d9642980\" (UID: \"21c59133-1783-4225-a70d-1395d9642980\") " Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.903715 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21c59133-1783-4225-a70d-1395d9642980-catalog-content\") pod \"21c59133-1783-4225-a70d-1395d9642980\" (UID: \"21c59133-1783-4225-a70d-1395d9642980\") " Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.903776 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffvww\" (UniqueName: \"kubernetes.io/projected/21c59133-1783-4225-a70d-1395d9642980-kube-api-access-ffvww\") pod \"21c59133-1783-4225-a70d-1395d9642980\" (UID: \"21c59133-1783-4225-a70d-1395d9642980\") " Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.904494 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21c59133-1783-4225-a70d-1395d9642980-utilities" (OuterVolumeSpecName: "utilities") pod "21c59133-1783-4225-a70d-1395d9642980" (UID: "21c59133-1783-4225-a70d-1395d9642980"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.906742 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21c59133-1783-4225-a70d-1395d9642980-kube-api-access-ffvww" (OuterVolumeSpecName: "kube-api-access-ffvww") pod "21c59133-1783-4225-a70d-1395d9642980" (UID: "21c59133-1783-4225-a70d-1395d9642980"). InnerVolumeSpecName "kube-api-access-ffvww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:58:39 crc kubenswrapper[4806]: E1204 03:58:39.926392 4806 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 04 03:58:39 crc kubenswrapper[4806]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-9rvth_openshift-marketplace_60ce8b7c-5cf6-452c-835e-9ab84a66aa65_0(f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad): error adding pod openshift-marketplace_marketplace-operator-79b997595-9rvth to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad" Netns:"/var/run/netns/12c0891c-b370-4078-a367-b4defa3c9738" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-9rvth;K8S_POD_INFRA_CONTAINER_ID=f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad;K8S_POD_UID=60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-9rvth] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-9rvth/60ce8b7c-5cf6-452c-835e-9ab84a66aa65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-9rvth?timeout=1m0s": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:58:39 crc kubenswrapper[4806]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 04 03:58:39 crc kubenswrapper[4806]: > Dec 04 03:58:39 crc kubenswrapper[4806]: E1204 03:58:39.926477 4806 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 04 03:58:39 crc kubenswrapper[4806]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-9rvth_openshift-marketplace_60ce8b7c-5cf6-452c-835e-9ab84a66aa65_0(f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad): error adding pod openshift-marketplace_marketplace-operator-79b997595-9rvth to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad" Netns:"/var/run/netns/12c0891c-b370-4078-a367-b4defa3c9738" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-9rvth;K8S_POD_INFRA_CONTAINER_ID=f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad;K8S_POD_UID=60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-9rvth] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-9rvth/60ce8b7c-5cf6-452c-835e-9ab84a66aa65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-9rvth?timeout=1m0s": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:58:39 crc kubenswrapper[4806]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 04 03:58:39 crc kubenswrapper[4806]: > pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:39 crc kubenswrapper[4806]: E1204 03:58:39.926503 4806 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 04 03:58:39 crc kubenswrapper[4806]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-9rvth_openshift-marketplace_60ce8b7c-5cf6-452c-835e-9ab84a66aa65_0(f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad): error adding pod openshift-marketplace_marketplace-operator-79b997595-9rvth to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad" Netns:"/var/run/netns/12c0891c-b370-4078-a367-b4defa3c9738" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-9rvth;K8S_POD_INFRA_CONTAINER_ID=f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad;K8S_POD_UID=60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-9rvth] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-9rvth/60ce8b7c-5cf6-452c-835e-9ab84a66aa65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-9rvth?timeout=1m0s": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:58:39 crc kubenswrapper[4806]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 04 03:58:39 crc kubenswrapper[4806]: > pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:39 crc kubenswrapper[4806]: E1204 03:58:39.926569 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"marketplace-operator-79b997595-9rvth_openshift-marketplace(60ce8b7c-5cf6-452c-835e-9ab84a66aa65)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"marketplace-operator-79b997595-9rvth_openshift-marketplace(60ce8b7c-5cf6-452c-835e-9ab84a66aa65)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-9rvth_openshift-marketplace_60ce8b7c-5cf6-452c-835e-9ab84a66aa65_0(f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad): error adding pod openshift-marketplace_marketplace-operator-79b997595-9rvth to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad\\\" Netns:\\\"/var/run/netns/12c0891c-b370-4078-a367-b4defa3c9738\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-9rvth;K8S_POD_INFRA_CONTAINER_ID=f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad;K8S_POD_UID=60ce8b7c-5cf6-452c-835e-9ab84a66aa65\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-9rvth] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-9rvth/60ce8b7c-5cf6-452c-835e-9ab84a66aa65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: status update failed for pod /: Get \\\"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-9rvth?timeout=1m0s\\\": dial tcp 38.102.83.17:6443: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" podUID="60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Dec 04 03:58:39 crc kubenswrapper[4806]: E1204 03:58:39.927521 4806 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.17:6443: connect: connection refused" event=< Dec 04 03:58:39 crc kubenswrapper[4806]: &Event{ObjectMeta:{marketplace-operator-79b997595-9rvth.187de71827782ff1 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:marketplace-operator-79b997595-9rvth,UID:60ce8b7c-5cf6-452c-835e-9ab84a66aa65,APIVersion:v1,ResourceVersion:29355,FieldPath:,},Reason:FailedCreatePodSandBox,Message:Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-9rvth_openshift-marketplace_60ce8b7c-5cf6-452c-835e-9ab84a66aa65_0(f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad): error adding pod openshift-marketplace_marketplace-operator-79b997595-9rvth to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad" Netns:"/var/run/netns/12c0891c-b370-4078-a367-b4defa3c9738" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-9rvth;K8S_POD_INFRA_CONTAINER_ID=f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad;K8S_POD_UID=60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-9rvth] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-9rvth/60ce8b7c-5cf6-452c-835e-9ab84a66aa65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-9rvth?timeout=1m0s": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:58:39 crc kubenswrapper[4806]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"},Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 03:58:39.926521841 +0000 UTC m=+234.785034759,LastTimestamp:2025-12-04 03:58:39.926521841 +0000 UTC m=+234.785034759,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 04 03:58:39 crc kubenswrapper[4806]: > Dec 04 03:58:39 crc kubenswrapper[4806]: I1204 03:58:39.989364 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21c59133-1783-4225-a70d-1395d9642980-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21c59133-1783-4225-a70d-1395d9642980" (UID: "21c59133-1783-4225-a70d-1395d9642980"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.004908 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21c59133-1783-4225-a70d-1395d9642980-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.004955 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21c59133-1783-4225-a70d-1395d9642980-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.004969 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffvww\" (UniqueName: \"kubernetes.io/projected/21c59133-1783-4225-a70d-1395d9642980-kube-api-access-ffvww\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.210376 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q9fqz" event={"ID":"59a2e27f-ed83-42ba-9371-14546fbcc6af","Type":"ContainerDied","Data":"f2807b5b634ebb72f89240c81f551a1a4cde072bf9a8bc29b22fab7acd4852ca"} Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.210433 4806 scope.go:117] "RemoveContainer" containerID="04dd0126cf689d5f7aa3f5021261660f9566c1317f784ebbc0a0e5b90f9a74df" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.210478 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q9fqz" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.213053 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlrnj" event={"ID":"21c59133-1783-4225-a70d-1395d9642980","Type":"ContainerDied","Data":"6d8e220fedeb7bcedbe609594e0264e951bb2760975027156d0b72c2118d7466"} Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.213244 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rlrnj" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.214594 4806 generic.go:334] "Generic (PLEG): container finished" podID="ed2b244e-28d2-406e-8b09-68e91494fa95" containerID="3982b9a8f94d1548d202054403e6d6a4da1b0311f261c078fe014badffcb4675" exitCode=0 Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.214650 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ed2b244e-28d2-406e-8b09-68e91494fa95","Type":"ContainerDied","Data":"3982b9a8f94d1548d202054403e6d6a4da1b0311f261c078fe014badffcb4675"} Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.216153 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" event={"ID":"4e4767a8-ff51-43c7-94cc-625c9b8689a3","Type":"ContainerDied","Data":"0d51f4cb252152e603656786b225cb1a33a5268ddddd3df2e38e9efff9648f5e"} Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.216163 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.219366 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ln2r5" event={"ID":"ef7e1243-9a57-4bbc-a28a-43cb93cf2add","Type":"ContainerDied","Data":"99116f9e4ba44a8de33bc4f00aaf26a6059a9a5d23db206cc189bc6b343fec9d"} Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.219440 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ln2r5" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.222307 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.222873 4806 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909" exitCode=0 Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.222899 4806 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9" exitCode=0 Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.222907 4806 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617" exitCode=0 Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.222915 4806 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44" exitCode=2 Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.226456 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.226483 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwdhx" event={"ID":"eaca0644-a305-41b3-84f1-4179e3412d37","Type":"ContainerDied","Data":"43424386062cf65cfae41f39a858e3613cdd4b0aaac1d9df3fb05ea8f3f1c555"} Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.226525 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwdhx" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.226835 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.229034 4806 scope.go:117] "RemoveContainer" containerID="769dff3547b6a3de75992d668dcbd705b51700b24d99cf13cb7499c2124acb1b" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.315004 4806 scope.go:117] "RemoveContainer" containerID="a41fbc6a04d874c280f0a8062014a00be43ae847613fb859d990c740824fe46b" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.347197 4806 scope.go:117] "RemoveContainer" containerID="84659a3af5d80e162ecbe638130986ac93820569328f6cb1a5f7a30d3d198440" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.384089 4806 scope.go:117] "RemoveContainer" containerID="27607f453517689111573accd547468b759759e2db407a8d687a6b85a5eee15d" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.403472 4806 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.403540 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.409332 4806 scope.go:117] "RemoveContainer" containerID="190950118eae2c27a4c54356561942af9725ab4f3cb228201e31e08d44905dd7" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.433487 4806 scope.go:117] "RemoveContainer" containerID="93979e497259d4fb735040c8bb60e0dbbf4d26aa5dad45718f9071972f3929ac" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.452392 4806 scope.go:117] "RemoveContainer" containerID="3d8a89e3e4eb267dcc1cb982fa7ce09d4b6fe168fad8c21218c0de647968c8f9" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.472852 4806 scope.go:117] "RemoveContainer" containerID="d9244d156597a2a7cdf05700dbc72398a240c92a12488f2daf0653099b1f4e8d" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.503263 4806 scope.go:117] "RemoveContainer" containerID="20071198a89a3299430557f0dcb1a2ae36ec656aaf903936cfdea7dcb15a09ed" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.528239 4806 scope.go:117] "RemoveContainer" containerID="aa4c2ceb9165d54155139b13b8e3dd58f12f8aa97ba2236faa3c276ca56eb542" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.548965 4806 scope.go:117] "RemoveContainer" containerID="32654e96eee457dd1b602189dfdb36283e390b07b7a62a8eae9f24d7b2ce8d7d" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.581490 4806 scope.go:117] "RemoveContainer" containerID="49e2a09d30e3e8f20dcfef1dd2669937c1f5416d41099e3686311653e8707fa3" Dec 04 03:58:40 crc kubenswrapper[4806]: E1204 03:58:40.675134 4806 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 04 03:58:40 crc kubenswrapper[4806]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-9rvth_openshift-marketplace_60ce8b7c-5cf6-452c-835e-9ab84a66aa65_0(39a6f38c6c03d1657b19ee534d2aed2cb04fdae34b52bfc508f8dafea423a5ea): error adding pod openshift-marketplace_marketplace-operator-79b997595-9rvth to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"39a6f38c6c03d1657b19ee534d2aed2cb04fdae34b52bfc508f8dafea423a5ea" Netns:"/var/run/netns/235ab4d0-00c1-42dd-a5ff-f1de9a2b8a6a" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-9rvth;K8S_POD_INFRA_CONTAINER_ID=39a6f38c6c03d1657b19ee534d2aed2cb04fdae34b52bfc508f8dafea423a5ea;K8S_POD_UID=60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-9rvth] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-9rvth/60ce8b7c-5cf6-452c-835e-9ab84a66aa65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-9rvth?timeout=1m0s": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:58:40 crc kubenswrapper[4806]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 04 03:58:40 crc kubenswrapper[4806]: > Dec 04 03:58:40 crc kubenswrapper[4806]: E1204 03:58:40.675201 4806 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 04 03:58:40 crc kubenswrapper[4806]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-9rvth_openshift-marketplace_60ce8b7c-5cf6-452c-835e-9ab84a66aa65_0(39a6f38c6c03d1657b19ee534d2aed2cb04fdae34b52bfc508f8dafea423a5ea): error adding pod openshift-marketplace_marketplace-operator-79b997595-9rvth to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"39a6f38c6c03d1657b19ee534d2aed2cb04fdae34b52bfc508f8dafea423a5ea" Netns:"/var/run/netns/235ab4d0-00c1-42dd-a5ff-f1de9a2b8a6a" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-9rvth;K8S_POD_INFRA_CONTAINER_ID=39a6f38c6c03d1657b19ee534d2aed2cb04fdae34b52bfc508f8dafea423a5ea;K8S_POD_UID=60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-9rvth] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-9rvth/60ce8b7c-5cf6-452c-835e-9ab84a66aa65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-9rvth?timeout=1m0s": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:58:40 crc kubenswrapper[4806]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 04 03:58:40 crc kubenswrapper[4806]: > pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:40 crc kubenswrapper[4806]: E1204 03:58:40.675224 4806 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 04 03:58:40 crc kubenswrapper[4806]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-9rvth_openshift-marketplace_60ce8b7c-5cf6-452c-835e-9ab84a66aa65_0(39a6f38c6c03d1657b19ee534d2aed2cb04fdae34b52bfc508f8dafea423a5ea): error adding pod openshift-marketplace_marketplace-operator-79b997595-9rvth to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"39a6f38c6c03d1657b19ee534d2aed2cb04fdae34b52bfc508f8dafea423a5ea" Netns:"/var/run/netns/235ab4d0-00c1-42dd-a5ff-f1de9a2b8a6a" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-9rvth;K8S_POD_INFRA_CONTAINER_ID=39a6f38c6c03d1657b19ee534d2aed2cb04fdae34b52bfc508f8dafea423a5ea;K8S_POD_UID=60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-9rvth] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-9rvth/60ce8b7c-5cf6-452c-835e-9ab84a66aa65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-9rvth?timeout=1m0s": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:58:40 crc kubenswrapper[4806]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 04 03:58:40 crc kubenswrapper[4806]: > pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:40 crc kubenswrapper[4806]: E1204 03:58:40.675276 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"marketplace-operator-79b997595-9rvth_openshift-marketplace(60ce8b7c-5cf6-452c-835e-9ab84a66aa65)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"marketplace-operator-79b997595-9rvth_openshift-marketplace(60ce8b7c-5cf6-452c-835e-9ab84a66aa65)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-9rvth_openshift-marketplace_60ce8b7c-5cf6-452c-835e-9ab84a66aa65_0(39a6f38c6c03d1657b19ee534d2aed2cb04fdae34b52bfc508f8dafea423a5ea): error adding pod openshift-marketplace_marketplace-operator-79b997595-9rvth to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"39a6f38c6c03d1657b19ee534d2aed2cb04fdae34b52bfc508f8dafea423a5ea\\\" Netns:\\\"/var/run/netns/235ab4d0-00c1-42dd-a5ff-f1de9a2b8a6a\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-9rvth;K8S_POD_INFRA_CONTAINER_ID=39a6f38c6c03d1657b19ee534d2aed2cb04fdae34b52bfc508f8dafea423a5ea;K8S_POD_UID=60ce8b7c-5cf6-452c-835e-9ab84a66aa65\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-9rvth] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-9rvth/60ce8b7c-5cf6-452c-835e-9ab84a66aa65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: status update failed for pod /: Get \\\"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-9rvth?timeout=1m0s\\\": dial tcp 38.102.83.17:6443: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" podUID="60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Dec 04 03:58:40 crc kubenswrapper[4806]: E1204 03:58:40.955906 4806 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:40 crc kubenswrapper[4806]: E1204 03:58:40.956585 4806 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:40 crc kubenswrapper[4806]: E1204 03:58:40.956813 4806 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:40 crc kubenswrapper[4806]: E1204 03:58:40.957016 4806 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:40 crc kubenswrapper[4806]: E1204 03:58:40.957194 4806 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:40 crc kubenswrapper[4806]: I1204 03:58:40.957221 4806 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 04 03:58:40 crc kubenswrapper[4806]: E1204 03:58:40.957380 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="200ms" Dec 04 03:58:41 crc kubenswrapper[4806]: E1204 03:58:41.158601 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="400ms" Dec 04 03:58:41 crc kubenswrapper[4806]: I1204 03:58:41.454694 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 03:58:41 crc kubenswrapper[4806]: I1204 03:58:41.524277 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ed2b244e-28d2-406e-8b09-68e91494fa95-var-lock\") pod \"ed2b244e-28d2-406e-8b09-68e91494fa95\" (UID: \"ed2b244e-28d2-406e-8b09-68e91494fa95\") " Dec 04 03:58:41 crc kubenswrapper[4806]: I1204 03:58:41.524381 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed2b244e-28d2-406e-8b09-68e91494fa95-kubelet-dir\") pod \"ed2b244e-28d2-406e-8b09-68e91494fa95\" (UID: \"ed2b244e-28d2-406e-8b09-68e91494fa95\") " Dec 04 03:58:41 crc kubenswrapper[4806]: I1204 03:58:41.524447 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed2b244e-28d2-406e-8b09-68e91494fa95-kube-api-access\") pod \"ed2b244e-28d2-406e-8b09-68e91494fa95\" (UID: \"ed2b244e-28d2-406e-8b09-68e91494fa95\") " Dec 04 03:58:41 crc kubenswrapper[4806]: I1204 03:58:41.525110 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed2b244e-28d2-406e-8b09-68e91494fa95-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ed2b244e-28d2-406e-8b09-68e91494fa95" (UID: "ed2b244e-28d2-406e-8b09-68e91494fa95"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 03:58:41 crc kubenswrapper[4806]: I1204 03:58:41.525111 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed2b244e-28d2-406e-8b09-68e91494fa95-var-lock" (OuterVolumeSpecName: "var-lock") pod "ed2b244e-28d2-406e-8b09-68e91494fa95" (UID: "ed2b244e-28d2-406e-8b09-68e91494fa95"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 03:58:41 crc kubenswrapper[4806]: I1204 03:58:41.531487 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed2b244e-28d2-406e-8b09-68e91494fa95-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ed2b244e-28d2-406e-8b09-68e91494fa95" (UID: "ed2b244e-28d2-406e-8b09-68e91494fa95"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 03:58:41 crc kubenswrapper[4806]: E1204 03:58:41.559705 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="800ms" Dec 04 03:58:41 crc kubenswrapper[4806]: I1204 03:58:41.626222 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ed2b244e-28d2-406e-8b09-68e91494fa95-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:41 crc kubenswrapper[4806]: I1204 03:58:41.626527 4806 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ed2b244e-28d2-406e-8b09-68e91494fa95-var-lock\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:41 crc kubenswrapper[4806]: I1204 03:58:41.626546 4806 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ed2b244e-28d2-406e-8b09-68e91494fa95-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:42 crc kubenswrapper[4806]: E1204 03:58:42.004131 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:58:41Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:58:41Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:58:41Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:58:41Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:42 crc kubenswrapper[4806]: E1204 03:58:42.004616 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:42 crc kubenswrapper[4806]: E1204 03:58:42.004815 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:42 crc kubenswrapper[4806]: E1204 03:58:42.004991 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:42 crc kubenswrapper[4806]: E1204 03:58:42.005440 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:42 crc kubenswrapper[4806]: E1204 03:58:42.005489 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 03:58:42 crc kubenswrapper[4806]: I1204 03:58:42.246504 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ed2b244e-28d2-406e-8b09-68e91494fa95","Type":"ContainerDied","Data":"e38cfe438a0976c9472417549ba39b4b16d19bc151f9419dbfe3fee4c7b0cada"} Dec 04 03:58:42 crc kubenswrapper[4806]: I1204 03:58:42.247488 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 03:58:42 crc kubenswrapper[4806]: I1204 03:58:42.247959 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e38cfe438a0976c9472417549ba39b4b16d19bc151f9419dbfe3fee4c7b0cada" Dec 04 03:58:42 crc kubenswrapper[4806]: I1204 03:58:42.253018 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 03:58:42 crc kubenswrapper[4806]: I1204 03:58:42.254625 4806 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da" exitCode=0 Dec 04 03:58:42 crc kubenswrapper[4806]: E1204 03:58:42.361051 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="1.6s" Dec 04 03:58:42 crc kubenswrapper[4806]: I1204 03:58:42.586463 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 03:58:42 crc kubenswrapper[4806]: I1204 03:58:42.587198 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:58:42 crc kubenswrapper[4806]: I1204 03:58:42.636969 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 03:58:42 crc kubenswrapper[4806]: I1204 03:58:42.637050 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 03:58:42 crc kubenswrapper[4806]: I1204 03:58:42.637105 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 03:58:42 crc kubenswrapper[4806]: I1204 03:58:42.637322 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 03:58:42 crc kubenswrapper[4806]: I1204 03:58:42.637353 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 03:58:42 crc kubenswrapper[4806]: I1204 03:58:42.637370 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 03:58:42 crc kubenswrapper[4806]: I1204 03:58:42.738724 4806 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:42 crc kubenswrapper[4806]: I1204 03:58:42.738775 4806 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:42 crc kubenswrapper[4806]: I1204 03:58:42.738794 4806 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 03:58:43 crc kubenswrapper[4806]: I1204 03:58:43.261990 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 03:58:43 crc kubenswrapper[4806]: I1204 03:58:43.262898 4806 scope.go:117] "RemoveContainer" containerID="1200d71cb8c21d3e0f02e749997dacfbb77409a6150a5ed34bfe05c1b7857909" Dec 04 03:58:43 crc kubenswrapper[4806]: I1204 03:58:43.263063 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:58:43 crc kubenswrapper[4806]: I1204 03:58:43.278870 4806 scope.go:117] "RemoveContainer" containerID="96e5710899ce3e1fbb74c47fff81e66f722f07827c908a0f0cdb2afaf80865b9" Dec 04 03:58:43 crc kubenswrapper[4806]: I1204 03:58:43.295916 4806 scope.go:117] "RemoveContainer" containerID="e74f5471e440c8c13932a0a09fc9731002bc27dd3dc8f615f9f2bca166e1c617" Dec 04 03:58:43 crc kubenswrapper[4806]: I1204 03:58:43.307463 4806 scope.go:117] "RemoveContainer" containerID="3167a452294bc0524432407cd3feb309a43e341910c81fbb5c17d1a282839b44" Dec 04 03:58:43 crc kubenswrapper[4806]: I1204 03:58:43.333005 4806 scope.go:117] "RemoveContainer" containerID="9ab6e6e7cb20ad677afd9695fc4e442e0343fa6c5f9c29df41b6d893fd6604da" Dec 04 03:58:43 crc kubenswrapper[4806]: I1204 03:58:43.359383 4806 scope.go:117] "RemoveContainer" containerID="7bf71975d2da84cc75a3e2ca3ccac5c34ddd27bcd3d15540784a9167f0209648" Dec 04 03:58:43 crc kubenswrapper[4806]: I1204 03:58:43.430628 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 04 03:58:43 crc kubenswrapper[4806]: E1204 03:58:43.961957 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="3.2s" Dec 04 03:58:44 crc kubenswrapper[4806]: I1204 03:58:44.511864 4806 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:44 crc kubenswrapper[4806]: I1204 03:58:44.512117 4806 status_manager.go:851] "Failed to get status for pod" podUID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-mzgjp\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:44 crc kubenswrapper[4806]: I1204 03:58:44.512366 4806 status_manager.go:851] "Failed to get status for pod" podUID="21c59133-1783-4225-a70d-1395d9642980" pod="openshift-marketplace/community-operators-rlrnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rlrnj\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:44 crc kubenswrapper[4806]: I1204 03:58:44.512694 4806 status_manager.go:851] "Failed to get status for pod" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" pod="openshift-marketplace/redhat-operators-hwdhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hwdhx\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:44 crc kubenswrapper[4806]: I1204 03:58:44.513002 4806 status_manager.go:851] "Failed to get status for pod" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" pod="openshift-marketplace/redhat-marketplace-ln2r5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ln2r5\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:44 crc kubenswrapper[4806]: I1204 03:58:44.513410 4806 status_manager.go:851] "Failed to get status for pod" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" pod="openshift-marketplace/certified-operators-q9fqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-q9fqz\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:44 crc kubenswrapper[4806]: I1204 03:58:44.513633 4806 status_manager.go:851] "Failed to get status for pod" podUID="ed2b244e-28d2-406e-8b09-68e91494fa95" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:44 crc kubenswrapper[4806]: I1204 03:58:44.513878 4806 status_manager.go:851] "Failed to get status for pod" podUID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-mzgjp\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:44 crc kubenswrapper[4806]: E1204 03:58:44.561904 4806 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:44 crc kubenswrapper[4806]: I1204 03:58:44.562211 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:45 crc kubenswrapper[4806]: I1204 03:58:45.276003 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"90c43b77c2e92b149636da0465d34af9cd0f36da9f964234b6dd25f4d2557193"} Dec 04 03:58:45 crc kubenswrapper[4806]: I1204 03:58:45.276356 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b10c45bb5b18b079a2279cd4e58573dbc16be3dde0d2c905a2ad0f177e2ebd05"} Dec 04 03:58:45 crc kubenswrapper[4806]: I1204 03:58:45.276908 4806 status_manager.go:851] "Failed to get status for pod" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" pod="openshift-marketplace/certified-operators-q9fqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-q9fqz\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:45 crc kubenswrapper[4806]: E1204 03:58:45.277046 4806 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:58:45 crc kubenswrapper[4806]: I1204 03:58:45.277363 4806 status_manager.go:851] "Failed to get status for pod" podUID="ed2b244e-28d2-406e-8b09-68e91494fa95" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:45 crc kubenswrapper[4806]: I1204 03:58:45.277735 4806 status_manager.go:851] "Failed to get status for pod" podUID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-mzgjp\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:45 crc kubenswrapper[4806]: I1204 03:58:45.278040 4806 status_manager.go:851] "Failed to get status for pod" podUID="21c59133-1783-4225-a70d-1395d9642980" pod="openshift-marketplace/community-operators-rlrnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rlrnj\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:45 crc kubenswrapper[4806]: I1204 03:58:45.278276 4806 status_manager.go:851] "Failed to get status for pod" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" pod="openshift-marketplace/redhat-operators-hwdhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hwdhx\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:45 crc kubenswrapper[4806]: I1204 03:58:45.278984 4806 status_manager.go:851] "Failed to get status for pod" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" pod="openshift-marketplace/redhat-marketplace-ln2r5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ln2r5\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:45 crc kubenswrapper[4806]: I1204 03:58:45.440668 4806 status_manager.go:851] "Failed to get status for pod" podUID="21c59133-1783-4225-a70d-1395d9642980" pod="openshift-marketplace/community-operators-rlrnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rlrnj\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:45 crc kubenswrapper[4806]: I1204 03:58:45.441416 4806 status_manager.go:851] "Failed to get status for pod" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" pod="openshift-marketplace/redhat-operators-hwdhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hwdhx\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:45 crc kubenswrapper[4806]: I1204 03:58:45.441708 4806 status_manager.go:851] "Failed to get status for pod" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" pod="openshift-marketplace/redhat-marketplace-ln2r5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ln2r5\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:45 crc kubenswrapper[4806]: I1204 03:58:45.442053 4806 status_manager.go:851] "Failed to get status for pod" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" pod="openshift-marketplace/certified-operators-q9fqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-q9fqz\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:45 crc kubenswrapper[4806]: I1204 03:58:45.442368 4806 status_manager.go:851] "Failed to get status for pod" podUID="ed2b244e-28d2-406e-8b09-68e91494fa95" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:45 crc kubenswrapper[4806]: I1204 03:58:45.443401 4806 status_manager.go:851] "Failed to get status for pod" podUID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-mzgjp\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:46 crc kubenswrapper[4806]: E1204 03:58:46.114813 4806 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.17:6443: connect: connection refused" event=< Dec 04 03:58:46 crc kubenswrapper[4806]: &Event{ObjectMeta:{marketplace-operator-79b997595-9rvth.187de71827782ff1 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:marketplace-operator-79b997595-9rvth,UID:60ce8b7c-5cf6-452c-835e-9ab84a66aa65,APIVersion:v1,ResourceVersion:29355,FieldPath:,},Reason:FailedCreatePodSandBox,Message:Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-9rvth_openshift-marketplace_60ce8b7c-5cf6-452c-835e-9ab84a66aa65_0(f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad): error adding pod openshift-marketplace_marketplace-operator-79b997595-9rvth to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad" Netns:"/var/run/netns/12c0891c-b370-4078-a367-b4defa3c9738" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-9rvth;K8S_POD_INFRA_CONTAINER_ID=f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad;K8S_POD_UID=60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-9rvth] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-9rvth/60ce8b7c-5cf6-452c-835e-9ab84a66aa65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-9rvth?timeout=1m0s": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:58:46 crc kubenswrapper[4806]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"},Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 03:58:39.926521841 +0000 UTC m=+234.785034759,LastTimestamp:2025-12-04 03:58:39.926521841 +0000 UTC m=+234.785034759,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 04 03:58:46 crc kubenswrapper[4806]: > Dec 04 03:58:47 crc kubenswrapper[4806]: E1204 03:58:47.163507 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="6.4s" Dec 04 03:58:52 crc kubenswrapper[4806]: E1204 03:58:52.098451 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:58:52Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:58:52Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:58:52Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T03:58:52Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:52 crc kubenswrapper[4806]: E1204 03:58:52.099432 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:52 crc kubenswrapper[4806]: E1204 03:58:52.099646 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:52 crc kubenswrapper[4806]: E1204 03:58:52.100032 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:52 crc kubenswrapper[4806]: E1204 03:58:52.100411 4806 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:52 crc kubenswrapper[4806]: E1204 03:58:52.100427 4806 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 03:58:52 crc kubenswrapper[4806]: E1204 03:58:52.473375 4806 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" volumeName="registry-storage" Dec 04 03:58:53 crc kubenswrapper[4806]: E1204 03:58:53.565004 4806 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="7s" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.051237 4806 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.051561 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.333229 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.333298 4806 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28" exitCode=1 Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.333334 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28"} Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.333775 4806 scope.go:117] "RemoveContainer" containerID="c84138c7f5f65939729dc446172c49eaa60d286183555433251655288d215e28" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.334082 4806 status_manager.go:851] "Failed to get status for pod" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" pod="openshift-marketplace/redhat-operators-hwdhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hwdhx\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.334576 4806 status_manager.go:851] "Failed to get status for pod" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" pod="openshift-marketplace/redhat-marketplace-ln2r5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ln2r5\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.334915 4806 status_manager.go:851] "Failed to get status for pod" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" pod="openshift-marketplace/certified-operators-q9fqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-q9fqz\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.335237 4806 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.335613 4806 status_manager.go:851] "Failed to get status for pod" podUID="ed2b244e-28d2-406e-8b09-68e91494fa95" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.335772 4806 status_manager.go:851] "Failed to get status for pod" podUID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-mzgjp\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.336066 4806 status_manager.go:851] "Failed to get status for pod" podUID="21c59133-1783-4225-a70d-1395d9642980" pod="openshift-marketplace/community-operators-rlrnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rlrnj\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.423684 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.427974 4806 status_manager.go:851] "Failed to get status for pod" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" pod="openshift-marketplace/redhat-marketplace-ln2r5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ln2r5\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.428699 4806 status_manager.go:851] "Failed to get status for pod" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" pod="openshift-marketplace/certified-operators-q9fqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-q9fqz\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.429199 4806 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.429674 4806 status_manager.go:851] "Failed to get status for pod" podUID="ed2b244e-28d2-406e-8b09-68e91494fa95" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.429990 4806 status_manager.go:851] "Failed to get status for pod" podUID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-mzgjp\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.430267 4806 status_manager.go:851] "Failed to get status for pod" podUID="21c59133-1783-4225-a70d-1395d9642980" pod="openshift-marketplace/community-operators-rlrnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rlrnj\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.430569 4806 status_manager.go:851] "Failed to get status for pod" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" pod="openshift-marketplace/redhat-operators-hwdhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hwdhx\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.431057 4806 status_manager.go:851] "Failed to get status for pod" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" pod="openshift-marketplace/certified-operators-q9fqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-q9fqz\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.431435 4806 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.432126 4806 status_manager.go:851] "Failed to get status for pod" podUID="ed2b244e-28d2-406e-8b09-68e91494fa95" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.432519 4806 status_manager.go:851] "Failed to get status for pod" podUID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-mzgjp\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.432910 4806 status_manager.go:851] "Failed to get status for pod" podUID="21c59133-1783-4225-a70d-1395d9642980" pod="openshift-marketplace/community-operators-rlrnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rlrnj\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.433139 4806 status_manager.go:851] "Failed to get status for pod" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" pod="openshift-marketplace/redhat-operators-hwdhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hwdhx\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.433405 4806 status_manager.go:851] "Failed to get status for pod" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" pod="openshift-marketplace/redhat-marketplace-ln2r5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ln2r5\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.446449 4806 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f45740b4-44fc-497c-9370-ad8b52fe52b4" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.446502 4806 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f45740b4-44fc-497c-9370-ad8b52fe52b4" Dec 04 03:58:55 crc kubenswrapper[4806]: E1204 03:58:55.447339 4806 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:58:55 crc kubenswrapper[4806]: I1204 03:58:55.447851 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:58:55 crc kubenswrapper[4806]: W1204 03:58:55.479011 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-12ab1501932577fb605fa2ea201c70a6f9b26aab7dbda62548232e07dd7b05f7 WatchSource:0}: Error finding container 12ab1501932577fb605fa2ea201c70a6f9b26aab7dbda62548232e07dd7b05f7: Status 404 returned error can't find the container with id 12ab1501932577fb605fa2ea201c70a6f9b26aab7dbda62548232e07dd7b05f7 Dec 04 03:58:56 crc kubenswrapper[4806]: E1204 03:58:56.115810 4806 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.17:6443: connect: connection refused" event=< Dec 04 03:58:56 crc kubenswrapper[4806]: &Event{ObjectMeta:{marketplace-operator-79b997595-9rvth.187de71827782ff1 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:marketplace-operator-79b997595-9rvth,UID:60ce8b7c-5cf6-452c-835e-9ab84a66aa65,APIVersion:v1,ResourceVersion:29355,FieldPath:,},Reason:FailedCreatePodSandBox,Message:Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-9rvth_openshift-marketplace_60ce8b7c-5cf6-452c-835e-9ab84a66aa65_0(f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad): error adding pod openshift-marketplace_marketplace-operator-79b997595-9rvth to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad" Netns:"/var/run/netns/12c0891c-b370-4078-a367-b4defa3c9738" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-9rvth;K8S_POD_INFRA_CONTAINER_ID=f0ae65f1079daf844380b4570043ca1690a0e654ee10a0693ca5929334f97cad;K8S_POD_UID=60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-9rvth] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-9rvth/60ce8b7c-5cf6-452c-835e-9ab84a66aa65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-9rvth in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-9rvth?timeout=1m0s": dial tcp 38.102.83.17:6443: connect: connection refused Dec 04 03:58:56 crc kubenswrapper[4806]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"},Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 03:58:39.926521841 +0000 UTC m=+234.785034759,LastTimestamp:2025-12-04 03:58:39.926521841 +0000 UTC m=+234.785034759,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 04 03:58:56 crc kubenswrapper[4806]: > Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.343555 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.343688 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c04ee26190a93a77b584b7a44a0ba77c3cad54437ee7b13787d852b86d050981"} Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.344589 4806 status_manager.go:851] "Failed to get status for pod" podUID="21c59133-1783-4225-a70d-1395d9642980" pod="openshift-marketplace/community-operators-rlrnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rlrnj\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.345030 4806 status_manager.go:851] "Failed to get status for pod" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" pod="openshift-marketplace/redhat-operators-hwdhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hwdhx\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.345255 4806 status_manager.go:851] "Failed to get status for pod" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" pod="openshift-marketplace/redhat-marketplace-ln2r5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ln2r5\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.345505 4806 status_manager.go:851] "Failed to get status for pod" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" pod="openshift-marketplace/certified-operators-q9fqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-q9fqz\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.345766 4806 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.346074 4806 status_manager.go:851] "Failed to get status for pod" podUID="ed2b244e-28d2-406e-8b09-68e91494fa95" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.346419 4806 status_manager.go:851] "Failed to get status for pod" podUID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-mzgjp\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.347419 4806 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="d440299b7488e5c2dbb6d735fbfa173adeae262d3ce7584882c7676e2e8fffc2" exitCode=0 Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.347457 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"d440299b7488e5c2dbb6d735fbfa173adeae262d3ce7584882c7676e2e8fffc2"} Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.347480 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"12ab1501932577fb605fa2ea201c70a6f9b26aab7dbda62548232e07dd7b05f7"} Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.347714 4806 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f45740b4-44fc-497c-9370-ad8b52fe52b4" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.347735 4806 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f45740b4-44fc-497c-9370-ad8b52fe52b4" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.347972 4806 status_manager.go:851] "Failed to get status for pod" podUID="21c59133-1783-4225-a70d-1395d9642980" pod="openshift-marketplace/community-operators-rlrnj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rlrnj\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:56 crc kubenswrapper[4806]: E1204 03:58:56.348097 4806 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.348292 4806 status_manager.go:851] "Failed to get status for pod" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" pod="openshift-marketplace/redhat-operators-hwdhx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-hwdhx\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.348606 4806 status_manager.go:851] "Failed to get status for pod" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" pod="openshift-marketplace/redhat-marketplace-ln2r5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ln2r5\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.348846 4806 status_manager.go:851] "Failed to get status for pod" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" pod="openshift-marketplace/certified-operators-q9fqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-q9fqz\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.349165 4806 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.349664 4806 status_manager.go:851] "Failed to get status for pod" podUID="ed2b244e-28d2-406e-8b09-68e91494fa95" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.349948 4806 status_manager.go:851] "Failed to get status for pod" podUID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" pod="openshift-marketplace/marketplace-operator-79b997595-mzgjp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-mzgjp\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.423451 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:56 crc kubenswrapper[4806]: I1204 03:58:56.424452 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:58:57 crc kubenswrapper[4806]: I1204 03:58:57.379768 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5451b80e19edc9c08d348760dde725e458f816695171e2a2c3cf350d54f7b151"} Dec 04 03:58:57 crc kubenswrapper[4806]: I1204 03:58:57.380416 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ef7e08a8a24404e0b75eb713dd1f16d8a6e7e63fee84ec123919556bcfe54c89"} Dec 04 03:58:57 crc kubenswrapper[4806]: I1204 03:58:57.380433 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0e8512fccff33dd5bdef919a94a6d262fba5836790215bbb1f3a5f976831a4e7"} Dec 04 03:58:57 crc kubenswrapper[4806]: I1204 03:58:57.380445 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e160398763789546dcc7e4126251ff2c682962dfac61c40f8fa201d724c6fb50"} Dec 04 03:58:58 crc kubenswrapper[4806]: I1204 03:58:58.387504 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fccd7b78316bc6552c29469abefa50b731b1a409b172984b257c0e7c1d255270"} Dec 04 03:58:58 crc kubenswrapper[4806]: I1204 03:58:58.387752 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:58:58 crc kubenswrapper[4806]: I1204 03:58:58.387881 4806 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f45740b4-44fc-497c-9370-ad8b52fe52b4" Dec 04 03:58:58 crc kubenswrapper[4806]: I1204 03:58:58.387914 4806 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f45740b4-44fc-497c-9370-ad8b52fe52b4" Dec 04 03:59:00 crc kubenswrapper[4806]: I1204 03:59:00.146971 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:59:00 crc kubenswrapper[4806]: I1204 03:59:00.151830 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:59:00 crc kubenswrapper[4806]: I1204 03:59:00.395641 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:59:00 crc kubenswrapper[4806]: I1204 03:59:00.448509 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:59:00 crc kubenswrapper[4806]: I1204 03:59:00.448599 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:59:00 crc kubenswrapper[4806]: I1204 03:59:00.452746 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:59:03 crc kubenswrapper[4806]: I1204 03:59:03.407656 4806 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:59:03 crc kubenswrapper[4806]: I1204 03:59:03.409344 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" event={"ID":"60ce8b7c-5cf6-452c-835e-9ab84a66aa65","Type":"ContainerStarted","Data":"7223a25b844d9790600abe6cb368a37729ff78b413db15fb78851ec5bb7e8cb2"} Dec 04 03:59:04 crc kubenswrapper[4806]: I1204 03:59:04.417467 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9rvth_60ce8b7c-5cf6-452c-835e-9ab84a66aa65/marketplace-operator/0.log" Dec 04 03:59:04 crc kubenswrapper[4806]: I1204 03:59:04.417538 4806 generic.go:334] "Generic (PLEG): container finished" podID="60ce8b7c-5cf6-452c-835e-9ab84a66aa65" containerID="896ec9580c3b3e2a663831a1fa006100ee205dec2548a202ed2ea117428a5c3c" exitCode=1 Dec 04 03:59:04 crc kubenswrapper[4806]: I1204 03:59:04.417641 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" event={"ID":"60ce8b7c-5cf6-452c-835e-9ab84a66aa65","Type":"ContainerDied","Data":"896ec9580c3b3e2a663831a1fa006100ee205dec2548a202ed2ea117428a5c3c"} Dec 04 03:59:04 crc kubenswrapper[4806]: I1204 03:59:04.418010 4806 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f45740b4-44fc-497c-9370-ad8b52fe52b4" Dec 04 03:59:04 crc kubenswrapper[4806]: I1204 03:59:04.418032 4806 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f45740b4-44fc-497c-9370-ad8b52fe52b4" Dec 04 03:59:04 crc kubenswrapper[4806]: I1204 03:59:04.418390 4806 scope.go:117] "RemoveContainer" containerID="896ec9580c3b3e2a663831a1fa006100ee205dec2548a202ed2ea117428a5c3c" Dec 04 03:59:04 crc kubenswrapper[4806]: I1204 03:59:04.423658 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:59:05 crc kubenswrapper[4806]: I1204 03:59:05.425057 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9rvth_60ce8b7c-5cf6-452c-835e-9ab84a66aa65/marketplace-operator/1.log" Dec 04 03:59:05 crc kubenswrapper[4806]: I1204 03:59:05.425770 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9rvth_60ce8b7c-5cf6-452c-835e-9ab84a66aa65/marketplace-operator/0.log" Dec 04 03:59:05 crc kubenswrapper[4806]: I1204 03:59:05.425808 4806 generic.go:334] "Generic (PLEG): container finished" podID="60ce8b7c-5cf6-452c-835e-9ab84a66aa65" containerID="dc372df064f2597b0f8c5cbd2de0e352c1511a0500d9d9a983bcbc39c292a4fd" exitCode=1 Dec 04 03:59:05 crc kubenswrapper[4806]: I1204 03:59:05.433864 4806 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f45740b4-44fc-497c-9370-ad8b52fe52b4" Dec 04 03:59:05 crc kubenswrapper[4806]: I1204 03:59:05.433899 4806 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f45740b4-44fc-497c-9370-ad8b52fe52b4" Dec 04 03:59:05 crc kubenswrapper[4806]: I1204 03:59:05.434014 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" event={"ID":"60ce8b7c-5cf6-452c-835e-9ab84a66aa65","Type":"ContainerDied","Data":"dc372df064f2597b0f8c5cbd2de0e352c1511a0500d9d9a983bcbc39c292a4fd"} Dec 04 03:59:05 crc kubenswrapper[4806]: I1204 03:59:05.434064 4806 scope.go:117] "RemoveContainer" containerID="896ec9580c3b3e2a663831a1fa006100ee205dec2548a202ed2ea117428a5c3c" Dec 04 03:59:05 crc kubenswrapper[4806]: I1204 03:59:05.434225 4806 scope.go:117] "RemoveContainer" containerID="dc372df064f2597b0f8c5cbd2de0e352c1511a0500d9d9a983bcbc39c292a4fd" Dec 04 03:59:05 crc kubenswrapper[4806]: E1204 03:59:05.434503 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-9rvth_openshift-marketplace(60ce8b7c-5cf6-452c-835e-9ab84a66aa65)\"" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" podUID="60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Dec 04 03:59:05 crc kubenswrapper[4806]: I1204 03:59:05.438427 4806 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="a27f504f-b022-41c1-a537-3f6cde5b2e0b" Dec 04 03:59:06 crc kubenswrapper[4806]: I1204 03:59:06.433668 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9rvth_60ce8b7c-5cf6-452c-835e-9ab84a66aa65/marketplace-operator/1.log" Dec 04 03:59:06 crc kubenswrapper[4806]: I1204 03:59:06.434627 4806 scope.go:117] "RemoveContainer" containerID="dc372df064f2597b0f8c5cbd2de0e352c1511a0500d9d9a983bcbc39c292a4fd" Dec 04 03:59:06 crc kubenswrapper[4806]: E1204 03:59:06.434948 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-9rvth_openshift-marketplace(60ce8b7c-5cf6-452c-835e-9ab84a66aa65)\"" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" podUID="60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Dec 04 03:59:09 crc kubenswrapper[4806]: I1204 03:59:09.217373 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:59:09 crc kubenswrapper[4806]: I1204 03:59:09.217750 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:59:09 crc kubenswrapper[4806]: I1204 03:59:09.218249 4806 scope.go:117] "RemoveContainer" containerID="dc372df064f2597b0f8c5cbd2de0e352c1511a0500d9d9a983bcbc39c292a4fd" Dec 04 03:59:09 crc kubenswrapper[4806]: E1204 03:59:09.218524 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-9rvth_openshift-marketplace(60ce8b7c-5cf6-452c-835e-9ab84a66aa65)\"" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" podUID="60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Dec 04 03:59:13 crc kubenswrapper[4806]: I1204 03:59:13.117724 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 04 03:59:13 crc kubenswrapper[4806]: I1204 03:59:13.328577 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 04 03:59:13 crc kubenswrapper[4806]: I1204 03:59:13.357810 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 03:59:14 crc kubenswrapper[4806]: I1204 03:59:14.152244 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 04 03:59:14 crc kubenswrapper[4806]: I1204 03:59:14.195336 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 04 03:59:14 crc kubenswrapper[4806]: I1204 03:59:14.482942 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 04 03:59:14 crc kubenswrapper[4806]: I1204 03:59:14.673951 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 04 03:59:14 crc kubenswrapper[4806]: I1204 03:59:14.798089 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 04 03:59:14 crc kubenswrapper[4806]: I1204 03:59:14.822612 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 04 03:59:15 crc kubenswrapper[4806]: I1204 03:59:15.030577 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 04 03:59:15 crc kubenswrapper[4806]: I1204 03:59:15.188971 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 04 03:59:15 crc kubenswrapper[4806]: I1204 03:59:15.238624 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 04 03:59:15 crc kubenswrapper[4806]: I1204 03:59:15.533846 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 04 03:59:15 crc kubenswrapper[4806]: I1204 03:59:15.545382 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 04 03:59:15 crc kubenswrapper[4806]: I1204 03:59:15.679252 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 04 03:59:15 crc kubenswrapper[4806]: I1204 03:59:15.711212 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 04 03:59:15 crc kubenswrapper[4806]: I1204 03:59:15.750566 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 04 03:59:15 crc kubenswrapper[4806]: I1204 03:59:15.961728 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 04 03:59:16 crc kubenswrapper[4806]: I1204 03:59:16.056349 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 04 03:59:16 crc kubenswrapper[4806]: I1204 03:59:16.056364 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 04 03:59:16 crc kubenswrapper[4806]: I1204 03:59:16.179778 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 04 03:59:16 crc kubenswrapper[4806]: I1204 03:59:16.239812 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 04 03:59:16 crc kubenswrapper[4806]: I1204 03:59:16.297829 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 04 03:59:16 crc kubenswrapper[4806]: I1204 03:59:16.321019 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 04 03:59:16 crc kubenswrapper[4806]: I1204 03:59:16.397133 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 04 03:59:16 crc kubenswrapper[4806]: I1204 03:59:16.583383 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 04 03:59:16 crc kubenswrapper[4806]: I1204 03:59:16.666637 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 04 03:59:16 crc kubenswrapper[4806]: I1204 03:59:16.839905 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.055478 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.149171 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.174153 4806 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.175589 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.209981 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.279104 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.329651 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.399154 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.444361 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.541420 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.579692 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.628632 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.664993 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.694180 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.777657 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.837994 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.838011 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.882454 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.908386 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.960164 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 04 03:59:17 crc kubenswrapper[4806]: I1204 03:59:17.963903 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.012189 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.022543 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.278402 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.312293 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.322469 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.366720 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.433339 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.620132 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.645070 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.677851 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.713472 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.740139 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.894367 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.913306 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.948365 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.980622 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 04 03:59:18 crc kubenswrapper[4806]: I1204 03:59:18.984099 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.006055 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.006355 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.071168 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.157098 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.203464 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.243599 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.321699 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.339571 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.429879 4806 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.498489 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.504074 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.513792 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.637880 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.644701 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.649468 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.900885 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 04 03:59:19 crc kubenswrapper[4806]: I1204 03:59:19.952547 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 03:59:20 crc kubenswrapper[4806]: I1204 03:59:20.019261 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 04 03:59:20 crc kubenswrapper[4806]: I1204 03:59:20.052998 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 04 03:59:20 crc kubenswrapper[4806]: I1204 03:59:20.072979 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 04 03:59:20 crc kubenswrapper[4806]: I1204 03:59:20.086183 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 04 03:59:20 crc kubenswrapper[4806]: I1204 03:59:20.127107 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 04 03:59:20 crc kubenswrapper[4806]: I1204 03:59:20.127146 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 04 03:59:20 crc kubenswrapper[4806]: I1204 03:59:20.582130 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 04 03:59:20 crc kubenswrapper[4806]: I1204 03:59:20.584351 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 04 03:59:20 crc kubenswrapper[4806]: I1204 03:59:20.642189 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 04 03:59:20 crc kubenswrapper[4806]: I1204 03:59:20.767477 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 04 03:59:20 crc kubenswrapper[4806]: I1204 03:59:20.776132 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 03:59:20 crc kubenswrapper[4806]: I1204 03:59:20.814036 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 04 03:59:20 crc kubenswrapper[4806]: I1204 03:59:20.854519 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 04 03:59:20 crc kubenswrapper[4806]: I1204 03:59:20.922706 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 04 03:59:21 crc kubenswrapper[4806]: I1204 03:59:21.100708 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 04 03:59:21 crc kubenswrapper[4806]: I1204 03:59:21.138896 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 04 03:59:21 crc kubenswrapper[4806]: I1204 03:59:21.182209 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 04 03:59:21 crc kubenswrapper[4806]: I1204 03:59:21.184055 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 04 03:59:21 crc kubenswrapper[4806]: I1204 03:59:21.212610 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 04 03:59:21 crc kubenswrapper[4806]: I1204 03:59:21.369859 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 04 03:59:21 crc kubenswrapper[4806]: I1204 03:59:21.424356 4806 scope.go:117] "RemoveContainer" containerID="dc372df064f2597b0f8c5cbd2de0e352c1511a0500d9d9a983bcbc39c292a4fd" Dec 04 03:59:21 crc kubenswrapper[4806]: I1204 03:59:21.478392 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 04 03:59:21 crc kubenswrapper[4806]: I1204 03:59:21.496236 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 04 03:59:21 crc kubenswrapper[4806]: I1204 03:59:21.527911 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 04 03:59:21 crc kubenswrapper[4806]: I1204 03:59:21.651431 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 04 03:59:21 crc kubenswrapper[4806]: I1204 03:59:21.868363 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 03:59:21 crc kubenswrapper[4806]: I1204 03:59:21.909094 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 04 03:59:21 crc kubenswrapper[4806]: I1204 03:59:21.933110 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 04 03:59:21 crc kubenswrapper[4806]: I1204 03:59:21.935734 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 04 03:59:21 crc kubenswrapper[4806]: I1204 03:59:21.969584 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.077637 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.087667 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.139827 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.196347 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.246407 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.270983 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.335450 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.366112 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.367596 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.404581 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.429556 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.464574 4806 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.532730 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9rvth_60ce8b7c-5cf6-452c-835e-9ab84a66aa65/marketplace-operator/2.log" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.533870 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9rvth_60ce8b7c-5cf6-452c-835e-9ab84a66aa65/marketplace-operator/1.log" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.533938 4806 generic.go:334] "Generic (PLEG): container finished" podID="60ce8b7c-5cf6-452c-835e-9ab84a66aa65" containerID="098a5c1b8b8e874934aac0bdf991a5f99125a34c7319fdbc142d06d867b1951f" exitCode=1 Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.533991 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" event={"ID":"60ce8b7c-5cf6-452c-835e-9ab84a66aa65","Type":"ContainerDied","Data":"098a5c1b8b8e874934aac0bdf991a5f99125a34c7319fdbc142d06d867b1951f"} Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.534050 4806 scope.go:117] "RemoveContainer" containerID="dc372df064f2597b0f8c5cbd2de0e352c1511a0500d9d9a983bcbc39c292a4fd" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.534806 4806 scope.go:117] "RemoveContainer" containerID="098a5c1b8b8e874934aac0bdf991a5f99125a34c7319fdbc142d06d867b1951f" Dec 04 03:59:22 crc kubenswrapper[4806]: E1204 03:59:22.535064 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-9rvth_openshift-marketplace(60ce8b7c-5cf6-452c-835e-9ab84a66aa65)\"" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" podUID="60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.587264 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.664222 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.665728 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.693897 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.843855 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.871783 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.925136 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 04 03:59:22 crc kubenswrapper[4806]: I1204 03:59:22.935821 4806 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.010785 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.242249 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.267144 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.312439 4806 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.314534 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.316429 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hwdhx","openshift-kube-apiserver/kube-apiserver-crc","openshift-marketplace/community-operators-rlrnj","openshift-marketplace/marketplace-operator-79b997595-mzgjp","openshift-marketplace/certified-operators-q9fqz","openshift-marketplace/redhat-marketplace-ln2r5"] Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.316509 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.316765 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9rvth"] Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.317128 4806 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f45740b4-44fc-497c-9370-ad8b52fe52b4" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.317184 4806 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f45740b4-44fc-497c-9370-ad8b52fe52b4" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.323146 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.339897 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=20.33986178 podStartE2EDuration="20.33986178s" podCreationTimestamp="2025-12-04 03:59:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:59:23.334862397 +0000 UTC m=+278.193375325" watchObservedRunningTime="2025-12-04 03:59:23.33986178 +0000 UTC m=+278.198374728" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.343427 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.383560 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.397793 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.429268 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.431799 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21c59133-1783-4225-a70d-1395d9642980" path="/var/lib/kubelet/pods/21c59133-1783-4225-a70d-1395d9642980/volumes" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.432656 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" path="/var/lib/kubelet/pods/4e4767a8-ff51-43c7-94cc-625c9b8689a3/volumes" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.433247 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" path="/var/lib/kubelet/pods/59a2e27f-ed83-42ba-9371-14546fbcc6af/volumes" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.434494 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaca0644-a305-41b3-84f1-4179e3412d37" path="/var/lib/kubelet/pods/eaca0644-a305-41b3-84f1-4179e3412d37/volumes" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.435194 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" path="/var/lib/kubelet/pods/ef7e1243-9a57-4bbc-a28a-43cb93cf2add/volumes" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.438041 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.472531 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.488421 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.515433 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.545215 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9rvth_60ce8b7c-5cf6-452c-835e-9ab84a66aa65/marketplace-operator/2.log" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.546103 4806 scope.go:117] "RemoveContainer" containerID="098a5c1b8b8e874934aac0bdf991a5f99125a34c7319fdbc142d06d867b1951f" Dec 04 03:59:23 crc kubenswrapper[4806]: E1204 03:59:23.546457 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-9rvth_openshift-marketplace(60ce8b7c-5cf6-452c-835e-9ab84a66aa65)\"" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" podUID="60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.577186 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.622706 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.625482 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.625588 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.634252 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.663876 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.685156 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.719752 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.789685 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 04 03:59:23 crc kubenswrapper[4806]: I1204 03:59:23.914078 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.092064 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.117347 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.293597 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.312449 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.327331 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.400186 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.569288 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.614202 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.637843 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.639032 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.704807 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.718370 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.777266 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.795250 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.824308 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.841818 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.899588 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.907780 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 04 03:59:24 crc kubenswrapper[4806]: I1204 03:59:24.933061 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 04 03:59:25 crc kubenswrapper[4806]: I1204 03:59:25.046088 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 04 03:59:25 crc kubenswrapper[4806]: I1204 03:59:25.200297 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 04 03:59:25 crc kubenswrapper[4806]: I1204 03:59:25.215832 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 04 03:59:25 crc kubenswrapper[4806]: I1204 03:59:25.230513 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 04 03:59:25 crc kubenswrapper[4806]: I1204 03:59:25.277383 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 04 03:59:25 crc kubenswrapper[4806]: I1204 03:59:25.285475 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 04 03:59:25 crc kubenswrapper[4806]: I1204 03:59:25.383007 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 04 03:59:25 crc kubenswrapper[4806]: I1204 03:59:25.453353 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 04 03:59:25 crc kubenswrapper[4806]: I1204 03:59:25.633895 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 04 03:59:25 crc kubenswrapper[4806]: I1204 03:59:25.642222 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 04 03:59:25 crc kubenswrapper[4806]: I1204 03:59:25.656894 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 04 03:59:25 crc kubenswrapper[4806]: I1204 03:59:25.766176 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 04 03:59:25 crc kubenswrapper[4806]: I1204 03:59:25.954590 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 04 03:59:25 crc kubenswrapper[4806]: I1204 03:59:25.979461 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 04 03:59:26 crc kubenswrapper[4806]: I1204 03:59:26.042898 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 04 03:59:26 crc kubenswrapper[4806]: I1204 03:59:26.077950 4806 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 03:59:26 crc kubenswrapper[4806]: I1204 03:59:26.078269 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://90c43b77c2e92b149636da0465d34af9cd0f36da9f964234b6dd25f4d2557193" gracePeriod=5 Dec 04 03:59:26 crc kubenswrapper[4806]: I1204 03:59:26.314039 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 04 03:59:26 crc kubenswrapper[4806]: I1204 03:59:26.408054 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 04 03:59:26 crc kubenswrapper[4806]: I1204 03:59:26.467797 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 04 03:59:26 crc kubenswrapper[4806]: I1204 03:59:26.536127 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 04 03:59:26 crc kubenswrapper[4806]: I1204 03:59:26.733968 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 04 03:59:26 crc kubenswrapper[4806]: I1204 03:59:26.746455 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 04 03:59:26 crc kubenswrapper[4806]: I1204 03:59:26.836760 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 04 03:59:26 crc kubenswrapper[4806]: I1204 03:59:26.849566 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 04 03:59:26 crc kubenswrapper[4806]: I1204 03:59:26.887681 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 04 03:59:27 crc kubenswrapper[4806]: I1204 03:59:27.286235 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 04 03:59:27 crc kubenswrapper[4806]: I1204 03:59:27.353632 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 04 03:59:27 crc kubenswrapper[4806]: I1204 03:59:27.363205 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 04 03:59:27 crc kubenswrapper[4806]: I1204 03:59:27.403345 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 04 03:59:27 crc kubenswrapper[4806]: I1204 03:59:27.454308 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 04 03:59:27 crc kubenswrapper[4806]: I1204 03:59:27.465972 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 04 03:59:27 crc kubenswrapper[4806]: I1204 03:59:27.585534 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 04 03:59:27 crc kubenswrapper[4806]: I1204 03:59:27.626361 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 04 03:59:27 crc kubenswrapper[4806]: I1204 03:59:27.630792 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 04 03:59:27 crc kubenswrapper[4806]: I1204 03:59:27.649616 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 04 03:59:27 crc kubenswrapper[4806]: I1204 03:59:27.826811 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 03:59:27 crc kubenswrapper[4806]: I1204 03:59:27.826826 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 04 03:59:27 crc kubenswrapper[4806]: I1204 03:59:27.924309 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 04 03:59:27 crc kubenswrapper[4806]: I1204 03:59:27.939971 4806 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 04 03:59:28 crc kubenswrapper[4806]: I1204 03:59:28.012391 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 04 03:59:28 crc kubenswrapper[4806]: I1204 03:59:28.018249 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 04 03:59:28 crc kubenswrapper[4806]: I1204 03:59:28.077247 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 04 03:59:28 crc kubenswrapper[4806]: I1204 03:59:28.222705 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 04 03:59:28 crc kubenswrapper[4806]: I1204 03:59:28.235125 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 03:59:28 crc kubenswrapper[4806]: I1204 03:59:28.601154 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 04 03:59:28 crc kubenswrapper[4806]: I1204 03:59:28.642361 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 04 03:59:28 crc kubenswrapper[4806]: I1204 03:59:28.667463 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 04 03:59:28 crc kubenswrapper[4806]: I1204 03:59:28.769698 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 04 03:59:28 crc kubenswrapper[4806]: I1204 03:59:28.861513 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 04 03:59:29 crc kubenswrapper[4806]: I1204 03:59:29.097176 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 04 03:59:29 crc kubenswrapper[4806]: I1204 03:59:29.217761 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:59:29 crc kubenswrapper[4806]: I1204 03:59:29.217831 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:59:29 crc kubenswrapper[4806]: I1204 03:59:29.218667 4806 scope.go:117] "RemoveContainer" containerID="098a5c1b8b8e874934aac0bdf991a5f99125a34c7319fdbc142d06d867b1951f" Dec 04 03:59:29 crc kubenswrapper[4806]: E1204 03:59:29.219057 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-9rvth_openshift-marketplace(60ce8b7c-5cf6-452c-835e-9ab84a66aa65)\"" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" podUID="60ce8b7c-5cf6-452c-835e-9ab84a66aa65" Dec 04 03:59:29 crc kubenswrapper[4806]: I1204 03:59:29.224349 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.608485 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.609211 4806 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="90c43b77c2e92b149636da0465d34af9cd0f36da9f964234b6dd25f4d2557193" exitCode=137 Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.668168 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.668275 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.845437 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.845616 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.845682 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.845757 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.845801 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.846213 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.846259 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.847420 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.847422 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.861108 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.947429 4806 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.947484 4806 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.947497 4806 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.947511 4806 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 04 03:59:31 crc kubenswrapper[4806]: I1204 03:59:31.947530 4806 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 04 03:59:32 crc kubenswrapper[4806]: I1204 03:59:32.619328 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 04 03:59:32 crc kubenswrapper[4806]: I1204 03:59:32.619446 4806 scope.go:117] "RemoveContainer" containerID="90c43b77c2e92b149636da0465d34af9cd0f36da9f964234b6dd25f4d2557193" Dec 04 03:59:32 crc kubenswrapper[4806]: I1204 03:59:32.619700 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 03:59:33 crc kubenswrapper[4806]: I1204 03:59:33.431909 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 04 03:59:40 crc kubenswrapper[4806]: I1204 03:59:40.764141 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 04 03:59:42 crc kubenswrapper[4806]: I1204 03:59:42.029731 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 04 03:59:42 crc kubenswrapper[4806]: I1204 03:59:42.367905 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 04 03:59:42 crc kubenswrapper[4806]: I1204 03:59:42.424267 4806 scope.go:117] "RemoveContainer" containerID="098a5c1b8b8e874934aac0bdf991a5f99125a34c7319fdbc142d06d867b1951f" Dec 04 03:59:42 crc kubenswrapper[4806]: I1204 03:59:42.690040 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9rvth_60ce8b7c-5cf6-452c-835e-9ab84a66aa65/marketplace-operator/2.log" Dec 04 03:59:42 crc kubenswrapper[4806]: I1204 03:59:42.690538 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" event={"ID":"60ce8b7c-5cf6-452c-835e-9ab84a66aa65","Type":"ContainerStarted","Data":"03bdf5061200c56bebf64416576a4a440995926c21a4edf17509839a6a433260"} Dec 04 03:59:42 crc kubenswrapper[4806]: I1204 03:59:42.690979 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:59:42 crc kubenswrapper[4806]: I1204 03:59:42.692448 4806 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9rvth container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" start-of-body= Dec 04 03:59:42 crc kubenswrapper[4806]: I1204 03:59:42.692515 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" podUID="60ce8b7c-5cf6-452c-835e-9ab84a66aa65" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" Dec 04 03:59:42 crc kubenswrapper[4806]: I1204 03:59:42.714300 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" podStartSLOduration=64.714276563 podStartE2EDuration="1m4.714276563s" podCreationTimestamp="2025-12-04 03:58:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 03:59:42.712388068 +0000 UTC m=+297.570900996" watchObservedRunningTime="2025-12-04 03:59:42.714276563 +0000 UTC m=+297.572789491" Dec 04 03:59:43 crc kubenswrapper[4806]: I1204 03:59:43.699512 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-9rvth" Dec 04 03:59:43 crc kubenswrapper[4806]: I1204 03:59:43.956681 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 03:59:45 crc kubenswrapper[4806]: I1204 03:59:45.740628 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 04 03:59:45 crc kubenswrapper[4806]: I1204 03:59:45.813028 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 04 03:59:46 crc kubenswrapper[4806]: I1204 03:59:46.808145 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 04 03:59:47 crc kubenswrapper[4806]: I1204 03:59:47.414223 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 04 03:59:48 crc kubenswrapper[4806]: I1204 03:59:48.225372 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 04 03:59:48 crc kubenswrapper[4806]: I1204 03:59:48.391277 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 04 03:59:48 crc kubenswrapper[4806]: I1204 03:59:48.769607 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 04 03:59:49 crc kubenswrapper[4806]: I1204 03:59:49.150797 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 04 03:59:49 crc kubenswrapper[4806]: I1204 03:59:49.686725 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 04 03:59:50 crc kubenswrapper[4806]: I1204 03:59:50.555818 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 04 03:59:53 crc kubenswrapper[4806]: I1204 03:59:53.495805 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 04 03:59:53 crc kubenswrapper[4806]: I1204 03:59:53.664897 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 04 03:59:57 crc kubenswrapper[4806]: I1204 03:59:57.838009 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.179731 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw"] Dec 04 04:00:00 crc kubenswrapper[4806]: E1204 04:00:00.180612 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21c59133-1783-4225-a70d-1395d9642980" containerName="registry-server" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.180635 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="21c59133-1783-4225-a70d-1395d9642980" containerName="registry-server" Dec 04 04:00:00 crc kubenswrapper[4806]: E1204 04:00:00.180650 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21c59133-1783-4225-a70d-1395d9642980" containerName="extract-content" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.180657 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="21c59133-1783-4225-a70d-1395d9642980" containerName="extract-content" Dec 04 04:00:00 crc kubenswrapper[4806]: E1204 04:00:00.180672 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" containerName="marketplace-operator" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.180680 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" containerName="marketplace-operator" Dec 04 04:00:00 crc kubenswrapper[4806]: E1204 04:00:00.180690 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21c59133-1783-4225-a70d-1395d9642980" containerName="extract-utilities" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.180697 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="21c59133-1783-4225-a70d-1395d9642980" containerName="extract-utilities" Dec 04 04:00:00 crc kubenswrapper[4806]: E1204 04:00:00.180710 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" containerName="registry-server" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.180717 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" containerName="registry-server" Dec 04 04:00:00 crc kubenswrapper[4806]: E1204 04:00:00.180728 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" containerName="extract-utilities" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.180735 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" containerName="extract-utilities" Dec 04 04:00:00 crc kubenswrapper[4806]: E1204 04:00:00.180749 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed2b244e-28d2-406e-8b09-68e91494fa95" containerName="installer" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.180755 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed2b244e-28d2-406e-8b09-68e91494fa95" containerName="installer" Dec 04 04:00:00 crc kubenswrapper[4806]: E1204 04:00:00.180768 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" containerName="extract-content" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.180776 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" containerName="extract-content" Dec 04 04:00:00 crc kubenswrapper[4806]: E1204 04:00:00.180795 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" containerName="extract-content" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.180802 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" containerName="extract-content" Dec 04 04:00:00 crc kubenswrapper[4806]: E1204 04:00:00.180811 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.180818 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 04:00:00 crc kubenswrapper[4806]: E1204 04:00:00.180831 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" containerName="registry-server" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.180838 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" containerName="registry-server" Dec 04 04:00:00 crc kubenswrapper[4806]: E1204 04:00:00.180850 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" containerName="extract-utilities" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.180860 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" containerName="extract-utilities" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.181006 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="21c59133-1783-4225-a70d-1395d9642980" containerName="registry-server" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.181026 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="59a2e27f-ed83-42ba-9371-14546fbcc6af" containerName="registry-server" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.181036 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed2b244e-28d2-406e-8b09-68e91494fa95" containerName="installer" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.181048 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef7e1243-9a57-4bbc-a28a-43cb93cf2add" containerName="registry-server" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.181064 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e4767a8-ff51-43c7-94cc-625c9b8689a3" containerName="marketplace-operator" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.181077 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.181735 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.186517 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.186882 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.197152 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw"] Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.383273 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-secret-volume\") pod \"collect-profiles-29413680-2hjnw\" (UID: \"d2da7dc1-86d1-4bbf-964d-9d9f83b17757\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.383347 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-config-volume\") pod \"collect-profiles-29413680-2hjnw\" (UID: \"d2da7dc1-86d1-4bbf-964d-9d9f83b17757\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.383408 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngsf8\" (UniqueName: \"kubernetes.io/projected/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-kube-api-access-ngsf8\") pod \"collect-profiles-29413680-2hjnw\" (UID: \"d2da7dc1-86d1-4bbf-964d-9d9f83b17757\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.485174 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-secret-volume\") pod \"collect-profiles-29413680-2hjnw\" (UID: \"d2da7dc1-86d1-4bbf-964d-9d9f83b17757\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.485855 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-config-volume\") pod \"collect-profiles-29413680-2hjnw\" (UID: \"d2da7dc1-86d1-4bbf-964d-9d9f83b17757\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.486037 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngsf8\" (UniqueName: \"kubernetes.io/projected/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-kube-api-access-ngsf8\") pod \"collect-profiles-29413680-2hjnw\" (UID: \"d2da7dc1-86d1-4bbf-964d-9d9f83b17757\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.488988 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-config-volume\") pod \"collect-profiles-29413680-2hjnw\" (UID: \"d2da7dc1-86d1-4bbf-964d-9d9f83b17757\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.494179 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-secret-volume\") pod \"collect-profiles-29413680-2hjnw\" (UID: \"d2da7dc1-86d1-4bbf-964d-9d9f83b17757\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.505239 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngsf8\" (UniqueName: \"kubernetes.io/projected/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-kube-api-access-ngsf8\") pod \"collect-profiles-29413680-2hjnw\" (UID: \"d2da7dc1-86d1-4bbf-964d-9d9f83b17757\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw" Dec 04 04:00:00 crc kubenswrapper[4806]: I1204 04:00:00.803655 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw" Dec 04 04:00:01 crc kubenswrapper[4806]: I1204 04:00:01.013863 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw"] Dec 04 04:00:01 crc kubenswrapper[4806]: I1204 04:00:01.495333 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 04 04:00:01 crc kubenswrapper[4806]: I1204 04:00:01.516199 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 04 04:00:01 crc kubenswrapper[4806]: I1204 04:00:01.807860 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 04 04:00:01 crc kubenswrapper[4806]: I1204 04:00:01.811905 4806 generic.go:334] "Generic (PLEG): container finished" podID="d2da7dc1-86d1-4bbf-964d-9d9f83b17757" containerID="34f58022124a4c35a0592a1f8fabe1e14b3fd4726b773fb9c61f91f509cfd15c" exitCode=0 Dec 04 04:00:01 crc kubenswrapper[4806]: I1204 04:00:01.812012 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw" event={"ID":"d2da7dc1-86d1-4bbf-964d-9d9f83b17757","Type":"ContainerDied","Data":"34f58022124a4c35a0592a1f8fabe1e14b3fd4726b773fb9c61f91f509cfd15c"} Dec 04 04:00:01 crc kubenswrapper[4806]: I1204 04:00:01.812067 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw" event={"ID":"d2da7dc1-86d1-4bbf-964d-9d9f83b17757","Type":"ContainerStarted","Data":"fb5b639843d6b88d5626e13af849395ced93fb0cf60bde2a5d27e695cae13ec5"} Dec 04 04:00:01 crc kubenswrapper[4806]: I1204 04:00:01.906380 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 04 04:00:03 crc kubenswrapper[4806]: I1204 04:00:03.044546 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw" Dec 04 04:00:03 crc kubenswrapper[4806]: I1204 04:00:03.226732 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-secret-volume\") pod \"d2da7dc1-86d1-4bbf-964d-9d9f83b17757\" (UID: \"d2da7dc1-86d1-4bbf-964d-9d9f83b17757\") " Dec 04 04:00:03 crc kubenswrapper[4806]: I1204 04:00:03.226784 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngsf8\" (UniqueName: \"kubernetes.io/projected/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-kube-api-access-ngsf8\") pod \"d2da7dc1-86d1-4bbf-964d-9d9f83b17757\" (UID: \"d2da7dc1-86d1-4bbf-964d-9d9f83b17757\") " Dec 04 04:00:03 crc kubenswrapper[4806]: I1204 04:00:03.226863 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-config-volume\") pod \"d2da7dc1-86d1-4bbf-964d-9d9f83b17757\" (UID: \"d2da7dc1-86d1-4bbf-964d-9d9f83b17757\") " Dec 04 04:00:03 crc kubenswrapper[4806]: I1204 04:00:03.227691 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-config-volume" (OuterVolumeSpecName: "config-volume") pod "d2da7dc1-86d1-4bbf-964d-9d9f83b17757" (UID: "d2da7dc1-86d1-4bbf-964d-9d9f83b17757"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:00:03 crc kubenswrapper[4806]: I1204 04:00:03.233763 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d2da7dc1-86d1-4bbf-964d-9d9f83b17757" (UID: "d2da7dc1-86d1-4bbf-964d-9d9f83b17757"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:00:03 crc kubenswrapper[4806]: I1204 04:00:03.235414 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-kube-api-access-ngsf8" (OuterVolumeSpecName: "kube-api-access-ngsf8") pod "d2da7dc1-86d1-4bbf-964d-9d9f83b17757" (UID: "d2da7dc1-86d1-4bbf-964d-9d9f83b17757"). InnerVolumeSpecName "kube-api-access-ngsf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:00:03 crc kubenswrapper[4806]: I1204 04:00:03.329006 4806 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:03 crc kubenswrapper[4806]: I1204 04:00:03.329109 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngsf8\" (UniqueName: \"kubernetes.io/projected/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-kube-api-access-ngsf8\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:03 crc kubenswrapper[4806]: I1204 04:00:03.329130 4806 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d2da7dc1-86d1-4bbf-964d-9d9f83b17757-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:03 crc kubenswrapper[4806]: I1204 04:00:03.826021 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw" event={"ID":"d2da7dc1-86d1-4bbf-964d-9d9f83b17757","Type":"ContainerDied","Data":"fb5b639843d6b88d5626e13af849395ced93fb0cf60bde2a5d27e695cae13ec5"} Dec 04 04:00:03 crc kubenswrapper[4806]: I1204 04:00:03.826099 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb5b639843d6b88d5626e13af849395ced93fb0cf60bde2a5d27e695cae13ec5" Dec 04 04:00:03 crc kubenswrapper[4806]: I1204 04:00:03.826092 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.441413 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zrtjm"] Dec 04 04:00:05 crc kubenswrapper[4806]: E1204 04:00:05.442575 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2da7dc1-86d1-4bbf-964d-9d9f83b17757" containerName="collect-profiles" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.442682 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2da7dc1-86d1-4bbf-964d-9d9f83b17757" containerName="collect-profiles" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.442935 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2da7dc1-86d1-4bbf-964d-9d9f83b17757" containerName="collect-profiles" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.444025 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zrtjm" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.457532 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.459031 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/242a6296-78ef-431a-8c26-522e66e5296c-catalog-content\") pod \"redhat-operators-zrtjm\" (UID: \"242a6296-78ef-431a-8c26-522e66e5296c\") " pod="openshift-marketplace/redhat-operators-zrtjm" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.459102 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdp6v\" (UniqueName: \"kubernetes.io/projected/242a6296-78ef-431a-8c26-522e66e5296c-kube-api-access-gdp6v\") pod \"redhat-operators-zrtjm\" (UID: \"242a6296-78ef-431a-8c26-522e66e5296c\") " pod="openshift-marketplace/redhat-operators-zrtjm" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.459133 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/242a6296-78ef-431a-8c26-522e66e5296c-utilities\") pod \"redhat-operators-zrtjm\" (UID: \"242a6296-78ef-431a-8c26-522e66e5296c\") " pod="openshift-marketplace/redhat-operators-zrtjm" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.459393 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zrtjm"] Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.488236 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.560725 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/242a6296-78ef-431a-8c26-522e66e5296c-catalog-content\") pod \"redhat-operators-zrtjm\" (UID: \"242a6296-78ef-431a-8c26-522e66e5296c\") " pod="openshift-marketplace/redhat-operators-zrtjm" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.560824 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdp6v\" (UniqueName: \"kubernetes.io/projected/242a6296-78ef-431a-8c26-522e66e5296c-kube-api-access-gdp6v\") pod \"redhat-operators-zrtjm\" (UID: \"242a6296-78ef-431a-8c26-522e66e5296c\") " pod="openshift-marketplace/redhat-operators-zrtjm" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.560868 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/242a6296-78ef-431a-8c26-522e66e5296c-utilities\") pod \"redhat-operators-zrtjm\" (UID: \"242a6296-78ef-431a-8c26-522e66e5296c\") " pod="openshift-marketplace/redhat-operators-zrtjm" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.561731 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/242a6296-78ef-431a-8c26-522e66e5296c-utilities\") pod \"redhat-operators-zrtjm\" (UID: \"242a6296-78ef-431a-8c26-522e66e5296c\") " pod="openshift-marketplace/redhat-operators-zrtjm" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.561719 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/242a6296-78ef-431a-8c26-522e66e5296c-catalog-content\") pod \"redhat-operators-zrtjm\" (UID: \"242a6296-78ef-431a-8c26-522e66e5296c\") " pod="openshift-marketplace/redhat-operators-zrtjm" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.580798 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdp6v\" (UniqueName: \"kubernetes.io/projected/242a6296-78ef-431a-8c26-522e66e5296c-kube-api-access-gdp6v\") pod \"redhat-operators-zrtjm\" (UID: \"242a6296-78ef-431a-8c26-522e66e5296c\") " pod="openshift-marketplace/redhat-operators-zrtjm" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.629537 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mkm6b"] Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.631044 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mkm6b" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.634029 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.646307 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mkm6b"] Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.663469 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3bcad8b-9a77-4d44-9bde-c3890fe6427c-catalog-content\") pod \"certified-operators-mkm6b\" (UID: \"c3bcad8b-9a77-4d44-9bde-c3890fe6427c\") " pod="openshift-marketplace/certified-operators-mkm6b" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.663528 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjszz\" (UniqueName: \"kubernetes.io/projected/c3bcad8b-9a77-4d44-9bde-c3890fe6427c-kube-api-access-jjszz\") pod \"certified-operators-mkm6b\" (UID: \"c3bcad8b-9a77-4d44-9bde-c3890fe6427c\") " pod="openshift-marketplace/certified-operators-mkm6b" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.663665 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3bcad8b-9a77-4d44-9bde-c3890fe6427c-utilities\") pod \"certified-operators-mkm6b\" (UID: \"c3bcad8b-9a77-4d44-9bde-c3890fe6427c\") " pod="openshift-marketplace/certified-operators-mkm6b" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.764583 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zrtjm" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.765416 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3bcad8b-9a77-4d44-9bde-c3890fe6427c-utilities\") pod \"certified-operators-mkm6b\" (UID: \"c3bcad8b-9a77-4d44-9bde-c3890fe6427c\") " pod="openshift-marketplace/certified-operators-mkm6b" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.765613 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3bcad8b-9a77-4d44-9bde-c3890fe6427c-catalog-content\") pod \"certified-operators-mkm6b\" (UID: \"c3bcad8b-9a77-4d44-9bde-c3890fe6427c\") " pod="openshift-marketplace/certified-operators-mkm6b" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.765759 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjszz\" (UniqueName: \"kubernetes.io/projected/c3bcad8b-9a77-4d44-9bde-c3890fe6427c-kube-api-access-jjszz\") pod \"certified-operators-mkm6b\" (UID: \"c3bcad8b-9a77-4d44-9bde-c3890fe6427c\") " pod="openshift-marketplace/certified-operators-mkm6b" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.766094 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3bcad8b-9a77-4d44-9bde-c3890fe6427c-catalog-content\") pod \"certified-operators-mkm6b\" (UID: \"c3bcad8b-9a77-4d44-9bde-c3890fe6427c\") " pod="openshift-marketplace/certified-operators-mkm6b" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.766402 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3bcad8b-9a77-4d44-9bde-c3890fe6427c-utilities\") pod \"certified-operators-mkm6b\" (UID: \"c3bcad8b-9a77-4d44-9bde-c3890fe6427c\") " pod="openshift-marketplace/certified-operators-mkm6b" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.796658 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjszz\" (UniqueName: \"kubernetes.io/projected/c3bcad8b-9a77-4d44-9bde-c3890fe6427c-kube-api-access-jjszz\") pod \"certified-operators-mkm6b\" (UID: \"c3bcad8b-9a77-4d44-9bde-c3890fe6427c\") " pod="openshift-marketplace/certified-operators-mkm6b" Dec 04 04:00:05 crc kubenswrapper[4806]: I1204 04:00:05.951474 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mkm6b" Dec 04 04:00:06 crc kubenswrapper[4806]: I1204 04:00:06.009484 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zrtjm"] Dec 04 04:00:06 crc kubenswrapper[4806]: I1204 04:00:06.200711 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mkm6b"] Dec 04 04:00:06 crc kubenswrapper[4806]: W1204 04:00:06.203143 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3bcad8b_9a77_4d44_9bde_c3890fe6427c.slice/crio-576e338da093e452f3eeab2298897dfdbc05419db1386ef3b753b08ad349bca4 WatchSource:0}: Error finding container 576e338da093e452f3eeab2298897dfdbc05419db1386ef3b753b08ad349bca4: Status 404 returned error can't find the container with id 576e338da093e452f3eeab2298897dfdbc05419db1386ef3b753b08ad349bca4 Dec 04 04:00:06 crc kubenswrapper[4806]: I1204 04:00:06.363851 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 04 04:00:06 crc kubenswrapper[4806]: I1204 04:00:06.846262 4806 generic.go:334] "Generic (PLEG): container finished" podID="242a6296-78ef-431a-8c26-522e66e5296c" containerID="36f71b0a21f9586e803509283bcc1ee63ad73fb2e3c6df4244a8af2453fc616b" exitCode=0 Dec 04 04:00:06 crc kubenswrapper[4806]: I1204 04:00:06.846502 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrtjm" event={"ID":"242a6296-78ef-431a-8c26-522e66e5296c","Type":"ContainerDied","Data":"36f71b0a21f9586e803509283bcc1ee63ad73fb2e3c6df4244a8af2453fc616b"} Dec 04 04:00:06 crc kubenswrapper[4806]: I1204 04:00:06.846553 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrtjm" event={"ID":"242a6296-78ef-431a-8c26-522e66e5296c","Type":"ContainerStarted","Data":"4f31c0621e081be3bf6d14c20e2beeed35385ab7939e32754eb66bab86215144"} Dec 04 04:00:06 crc kubenswrapper[4806]: I1204 04:00:06.854620 4806 generic.go:334] "Generic (PLEG): container finished" podID="c3bcad8b-9a77-4d44-9bde-c3890fe6427c" containerID="0f0a947e8276fc18340eed1612fd49a4f83cd17f3352b0a3e7c0874de1959752" exitCode=0 Dec 04 04:00:06 crc kubenswrapper[4806]: I1204 04:00:06.854693 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkm6b" event={"ID":"c3bcad8b-9a77-4d44-9bde-c3890fe6427c","Type":"ContainerDied","Data":"0f0a947e8276fc18340eed1612fd49a4f83cd17f3352b0a3e7c0874de1959752"} Dec 04 04:00:06 crc kubenswrapper[4806]: I1204 04:00:06.854767 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkm6b" event={"ID":"c3bcad8b-9a77-4d44-9bde-c3890fe6427c","Type":"ContainerStarted","Data":"576e338da093e452f3eeab2298897dfdbc05419db1386ef3b753b08ad349bca4"} Dec 04 04:00:07 crc kubenswrapper[4806]: I1204 04:00:07.615107 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 04 04:00:07 crc kubenswrapper[4806]: I1204 04:00:07.830876 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2bzpg"] Dec 04 04:00:07 crc kubenswrapper[4806]: I1204 04:00:07.832413 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2bzpg" Dec 04 04:00:07 crc kubenswrapper[4806]: I1204 04:00:07.835263 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 04:00:07 crc kubenswrapper[4806]: I1204 04:00:07.852097 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2bzpg"] Dec 04 04:00:07 crc kubenswrapper[4806]: I1204 04:00:07.862637 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkm6b" event={"ID":"c3bcad8b-9a77-4d44-9bde-c3890fe6427c","Type":"ContainerStarted","Data":"2e726d06df00346a31f27e7597eebb72b4964ae903ce4db0cc1fb0bc73827d65"} Dec 04 04:00:07 crc kubenswrapper[4806]: I1204 04:00:07.863900 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrtjm" event={"ID":"242a6296-78ef-431a-8c26-522e66e5296c","Type":"ContainerStarted","Data":"c4319011c9b7c8927a8a8773189ad8c7da985c9f2342f47dd07977f235d5d956"} Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.018650 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bd6d288-1d0e-49a2-b9de-5437c5d22356-catalog-content\") pod \"community-operators-2bzpg\" (UID: \"0bd6d288-1d0e-49a2-b9de-5437c5d22356\") " pod="openshift-marketplace/community-operators-2bzpg" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.018798 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6jm8\" (UniqueName: \"kubernetes.io/projected/0bd6d288-1d0e-49a2-b9de-5437c5d22356-kube-api-access-z6jm8\") pod \"community-operators-2bzpg\" (UID: \"0bd6d288-1d0e-49a2-b9de-5437c5d22356\") " pod="openshift-marketplace/community-operators-2bzpg" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.018835 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bd6d288-1d0e-49a2-b9de-5437c5d22356-utilities\") pod \"community-operators-2bzpg\" (UID: \"0bd6d288-1d0e-49a2-b9de-5437c5d22356\") " pod="openshift-marketplace/community-operators-2bzpg" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.120473 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bd6d288-1d0e-49a2-b9de-5437c5d22356-catalog-content\") pod \"community-operators-2bzpg\" (UID: \"0bd6d288-1d0e-49a2-b9de-5437c5d22356\") " pod="openshift-marketplace/community-operators-2bzpg" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.120615 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6jm8\" (UniqueName: \"kubernetes.io/projected/0bd6d288-1d0e-49a2-b9de-5437c5d22356-kube-api-access-z6jm8\") pod \"community-operators-2bzpg\" (UID: \"0bd6d288-1d0e-49a2-b9de-5437c5d22356\") " pod="openshift-marketplace/community-operators-2bzpg" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.120662 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bd6d288-1d0e-49a2-b9de-5437c5d22356-utilities\") pod \"community-operators-2bzpg\" (UID: \"0bd6d288-1d0e-49a2-b9de-5437c5d22356\") " pod="openshift-marketplace/community-operators-2bzpg" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.121170 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bd6d288-1d0e-49a2-b9de-5437c5d22356-catalog-content\") pod \"community-operators-2bzpg\" (UID: \"0bd6d288-1d0e-49a2-b9de-5437c5d22356\") " pod="openshift-marketplace/community-operators-2bzpg" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.121230 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bd6d288-1d0e-49a2-b9de-5437c5d22356-utilities\") pod \"community-operators-2bzpg\" (UID: \"0bd6d288-1d0e-49a2-b9de-5437c5d22356\") " pod="openshift-marketplace/community-operators-2bzpg" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.145002 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6jm8\" (UniqueName: \"kubernetes.io/projected/0bd6d288-1d0e-49a2-b9de-5437c5d22356-kube-api-access-z6jm8\") pod \"community-operators-2bzpg\" (UID: \"0bd6d288-1d0e-49a2-b9de-5437c5d22356\") " pod="openshift-marketplace/community-operators-2bzpg" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.152136 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2bzpg" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.230756 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kgr4l"] Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.234347 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kgr4l" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.237053 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.241472 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kgr4l"] Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.365071 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2bzpg"] Dec 04 04:00:08 crc kubenswrapper[4806]: W1204 04:00:08.381258 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0bd6d288_1d0e_49a2_b9de_5437c5d22356.slice/crio-4e04b248ffbfa56bf05af605cf1bfa081dc9318f3afa033b6f70523f9de6c2fc WatchSource:0}: Error finding container 4e04b248ffbfa56bf05af605cf1bfa081dc9318f3afa033b6f70523f9de6c2fc: Status 404 returned error can't find the container with id 4e04b248ffbfa56bf05af605cf1bfa081dc9318f3afa033b6f70523f9de6c2fc Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.425389 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s588z\" (UniqueName: \"kubernetes.io/projected/7bf8469a-647b-4b61-ab6c-89bbcc436ab9-kube-api-access-s588z\") pod \"redhat-marketplace-kgr4l\" (UID: \"7bf8469a-647b-4b61-ab6c-89bbcc436ab9\") " pod="openshift-marketplace/redhat-marketplace-kgr4l" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.425501 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bf8469a-647b-4b61-ab6c-89bbcc436ab9-utilities\") pod \"redhat-marketplace-kgr4l\" (UID: \"7bf8469a-647b-4b61-ab6c-89bbcc436ab9\") " pod="openshift-marketplace/redhat-marketplace-kgr4l" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.425556 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bf8469a-647b-4b61-ab6c-89bbcc436ab9-catalog-content\") pod \"redhat-marketplace-kgr4l\" (UID: \"7bf8469a-647b-4b61-ab6c-89bbcc436ab9\") " pod="openshift-marketplace/redhat-marketplace-kgr4l" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.527293 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bf8469a-647b-4b61-ab6c-89bbcc436ab9-utilities\") pod \"redhat-marketplace-kgr4l\" (UID: \"7bf8469a-647b-4b61-ab6c-89bbcc436ab9\") " pod="openshift-marketplace/redhat-marketplace-kgr4l" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.527954 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bf8469a-647b-4b61-ab6c-89bbcc436ab9-catalog-content\") pod \"redhat-marketplace-kgr4l\" (UID: \"7bf8469a-647b-4b61-ab6c-89bbcc436ab9\") " pod="openshift-marketplace/redhat-marketplace-kgr4l" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.527992 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s588z\" (UniqueName: \"kubernetes.io/projected/7bf8469a-647b-4b61-ab6c-89bbcc436ab9-kube-api-access-s588z\") pod \"redhat-marketplace-kgr4l\" (UID: \"7bf8469a-647b-4b61-ab6c-89bbcc436ab9\") " pod="openshift-marketplace/redhat-marketplace-kgr4l" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.528126 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bf8469a-647b-4b61-ab6c-89bbcc436ab9-utilities\") pod \"redhat-marketplace-kgr4l\" (UID: \"7bf8469a-647b-4b61-ab6c-89bbcc436ab9\") " pod="openshift-marketplace/redhat-marketplace-kgr4l" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.528678 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bf8469a-647b-4b61-ab6c-89bbcc436ab9-catalog-content\") pod \"redhat-marketplace-kgr4l\" (UID: \"7bf8469a-647b-4b61-ab6c-89bbcc436ab9\") " pod="openshift-marketplace/redhat-marketplace-kgr4l" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.552608 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s588z\" (UniqueName: \"kubernetes.io/projected/7bf8469a-647b-4b61-ab6c-89bbcc436ab9-kube-api-access-s588z\") pod \"redhat-marketplace-kgr4l\" (UID: \"7bf8469a-647b-4b61-ab6c-89bbcc436ab9\") " pod="openshift-marketplace/redhat-marketplace-kgr4l" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.587639 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kgr4l" Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.793100 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kgr4l"] Dec 04 04:00:08 crc kubenswrapper[4806]: W1204 04:00:08.801462 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bf8469a_647b_4b61_ab6c_89bbcc436ab9.slice/crio-059604d6e0ab33afd6ebe258cfca538ffa8a68b2583700578a736b21b9218776 WatchSource:0}: Error finding container 059604d6e0ab33afd6ebe258cfca538ffa8a68b2583700578a736b21b9218776: Status 404 returned error can't find the container with id 059604d6e0ab33afd6ebe258cfca538ffa8a68b2583700578a736b21b9218776 Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.873357 4806 generic.go:334] "Generic (PLEG): container finished" podID="c3bcad8b-9a77-4d44-9bde-c3890fe6427c" containerID="2e726d06df00346a31f27e7597eebb72b4964ae903ce4db0cc1fb0bc73827d65" exitCode=0 Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.874025 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkm6b" event={"ID":"c3bcad8b-9a77-4d44-9bde-c3890fe6427c","Type":"ContainerDied","Data":"2e726d06df00346a31f27e7597eebb72b4964ae903ce4db0cc1fb0bc73827d65"} Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.880345 4806 generic.go:334] "Generic (PLEG): container finished" podID="242a6296-78ef-431a-8c26-522e66e5296c" containerID="c4319011c9b7c8927a8a8773189ad8c7da985c9f2342f47dd07977f235d5d956" exitCode=0 Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.880414 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrtjm" event={"ID":"242a6296-78ef-431a-8c26-522e66e5296c","Type":"ContainerDied","Data":"c4319011c9b7c8927a8a8773189ad8c7da985c9f2342f47dd07977f235d5d956"} Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.882810 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgr4l" event={"ID":"7bf8469a-647b-4b61-ab6c-89bbcc436ab9","Type":"ContainerStarted","Data":"059604d6e0ab33afd6ebe258cfca538ffa8a68b2583700578a736b21b9218776"} Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.887670 4806 generic.go:334] "Generic (PLEG): container finished" podID="0bd6d288-1d0e-49a2-b9de-5437c5d22356" containerID="3cc3b3f216c51d28ba750f108f805e59e13d06a97fa22374cfa7fe54f063e152" exitCode=0 Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.887714 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2bzpg" event={"ID":"0bd6d288-1d0e-49a2-b9de-5437c5d22356","Type":"ContainerDied","Data":"3cc3b3f216c51d28ba750f108f805e59e13d06a97fa22374cfa7fe54f063e152"} Dec 04 04:00:08 crc kubenswrapper[4806]: I1204 04:00:08.887748 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2bzpg" event={"ID":"0bd6d288-1d0e-49a2-b9de-5437c5d22356","Type":"ContainerStarted","Data":"4e04b248ffbfa56bf05af605cf1bfa081dc9318f3afa033b6f70523f9de6c2fc"} Dec 04 04:00:09 crc kubenswrapper[4806]: I1204 04:00:09.896466 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkm6b" event={"ID":"c3bcad8b-9a77-4d44-9bde-c3890fe6427c","Type":"ContainerStarted","Data":"ad69963a7938a86520f92313c27423b2a185fd806c5aef3aef45d3566c338633"} Dec 04 04:00:09 crc kubenswrapper[4806]: I1204 04:00:09.903350 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zrtjm" event={"ID":"242a6296-78ef-431a-8c26-522e66e5296c","Type":"ContainerStarted","Data":"9681065033be8e73907fa56fe2eacff11a67c6955372d8b052996a1e5b653d8e"} Dec 04 04:00:09 crc kubenswrapper[4806]: I1204 04:00:09.904818 4806 generic.go:334] "Generic (PLEG): container finished" podID="7bf8469a-647b-4b61-ab6c-89bbcc436ab9" containerID="67328d401f3ca07cdc1685a14609c70fec37584a5ea0a6f8fdfe08474f019614" exitCode=0 Dec 04 04:00:09 crc kubenswrapper[4806]: I1204 04:00:09.904889 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgr4l" event={"ID":"7bf8469a-647b-4b61-ab6c-89bbcc436ab9","Type":"ContainerDied","Data":"67328d401f3ca07cdc1685a14609c70fec37584a5ea0a6f8fdfe08474f019614"} Dec 04 04:00:09 crc kubenswrapper[4806]: I1204 04:00:09.908196 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2bzpg" event={"ID":"0bd6d288-1d0e-49a2-b9de-5437c5d22356","Type":"ContainerStarted","Data":"eb9ad0126cff4080243369c10aa94014ad48993d7082894ef8b55a1a30fbe2bc"} Dec 04 04:00:09 crc kubenswrapper[4806]: I1204 04:00:09.929291 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mkm6b" podStartSLOduration=2.471075705 podStartE2EDuration="4.929226296s" podCreationTimestamp="2025-12-04 04:00:05 +0000 UTC" firstStartedPulling="2025-12-04 04:00:06.858438375 +0000 UTC m=+321.716951303" lastFinishedPulling="2025-12-04 04:00:09.316588966 +0000 UTC m=+324.175101894" observedRunningTime="2025-12-04 04:00:09.926228822 +0000 UTC m=+324.784741770" watchObservedRunningTime="2025-12-04 04:00:09.929226296 +0000 UTC m=+324.787739224" Dec 04 04:00:10 crc kubenswrapper[4806]: I1204 04:00:10.012356 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zrtjm" podStartSLOduration=2.348120971 podStartE2EDuration="5.012326339s" podCreationTimestamp="2025-12-04 04:00:05 +0000 UTC" firstStartedPulling="2025-12-04 04:00:06.84932869 +0000 UTC m=+321.707841618" lastFinishedPulling="2025-12-04 04:00:09.513534068 +0000 UTC m=+324.372046986" observedRunningTime="2025-12-04 04:00:10.008719453 +0000 UTC m=+324.867232391" watchObservedRunningTime="2025-12-04 04:00:10.012326339 +0000 UTC m=+324.870839267" Dec 04 04:00:10 crc kubenswrapper[4806]: I1204 04:00:10.917262 4806 generic.go:334] "Generic (PLEG): container finished" podID="0bd6d288-1d0e-49a2-b9de-5437c5d22356" containerID="eb9ad0126cff4080243369c10aa94014ad48993d7082894ef8b55a1a30fbe2bc" exitCode=0 Dec 04 04:00:10 crc kubenswrapper[4806]: I1204 04:00:10.917337 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2bzpg" event={"ID":"0bd6d288-1d0e-49a2-b9de-5437c5d22356","Type":"ContainerDied","Data":"eb9ad0126cff4080243369c10aa94014ad48993d7082894ef8b55a1a30fbe2bc"} Dec 04 04:00:10 crc kubenswrapper[4806]: I1204 04:00:10.922154 4806 generic.go:334] "Generic (PLEG): container finished" podID="7bf8469a-647b-4b61-ab6c-89bbcc436ab9" containerID="80a953fb48d21b661b5b8a649bdb5a0f0abfff40c12cc723698e6c04c5ac3edc" exitCode=0 Dec 04 04:00:10 crc kubenswrapper[4806]: I1204 04:00:10.923769 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgr4l" event={"ID":"7bf8469a-647b-4b61-ab6c-89bbcc436ab9","Type":"ContainerDied","Data":"80a953fb48d21b661b5b8a649bdb5a0f0abfff40c12cc723698e6c04c5ac3edc"} Dec 04 04:00:11 crc kubenswrapper[4806]: I1204 04:00:11.446352 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 04 04:00:11 crc kubenswrapper[4806]: I1204 04:00:11.930661 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kgr4l" event={"ID":"7bf8469a-647b-4b61-ab6c-89bbcc436ab9","Type":"ContainerStarted","Data":"5702ae29df3be71a62e0b1966ace208a7bc135a15aad970c392084172757f965"} Dec 04 04:00:11 crc kubenswrapper[4806]: I1204 04:00:11.933973 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2bzpg" event={"ID":"0bd6d288-1d0e-49a2-b9de-5437c5d22356","Type":"ContainerStarted","Data":"750efff64f82e0e8ed73381fff0c5fd0c4fd97f43a6ad514d4496d84ec9322cf"} Dec 04 04:00:11 crc kubenswrapper[4806]: I1204 04:00:11.963371 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kgr4l" podStartSLOduration=2.545166841 podStartE2EDuration="3.96334212s" podCreationTimestamp="2025-12-04 04:00:08 +0000 UTC" firstStartedPulling="2025-12-04 04:00:09.906100434 +0000 UTC m=+324.764613362" lastFinishedPulling="2025-12-04 04:00:11.324275723 +0000 UTC m=+326.182788641" observedRunningTime="2025-12-04 04:00:11.961347101 +0000 UTC m=+326.819860029" watchObservedRunningTime="2025-12-04 04:00:11.96334212 +0000 UTC m=+326.821855048" Dec 04 04:00:11 crc kubenswrapper[4806]: I1204 04:00:11.989768 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2bzpg" podStartSLOduration=2.500525936 podStartE2EDuration="4.989739785s" podCreationTimestamp="2025-12-04 04:00:07 +0000 UTC" firstStartedPulling="2025-12-04 04:00:08.889299876 +0000 UTC m=+323.747812804" lastFinishedPulling="2025-12-04 04:00:11.378513725 +0000 UTC m=+326.237026653" observedRunningTime="2025-12-04 04:00:11.987151155 +0000 UTC m=+326.845664103" watchObservedRunningTime="2025-12-04 04:00:11.989739785 +0000 UTC m=+326.848252733" Dec 04 04:00:14 crc kubenswrapper[4806]: I1204 04:00:14.909381 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pddk8"] Dec 04 04:00:14 crc kubenswrapper[4806]: I1204 04:00:14.910435 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" podUID="606e54dc-5cc4-4e02-b796-dd37b613a8d0" containerName="controller-manager" containerID="cri-o://6be8b06358583e17b0687a963e9bb13718431db5a968e305e3eb98c49e5ce081" gracePeriod=30 Dec 04 04:00:15 crc kubenswrapper[4806]: I1204 04:00:15.007878 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml"] Dec 04 04:00:15 crc kubenswrapper[4806]: I1204 04:00:15.008227 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" podUID="735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8" containerName="route-controller-manager" containerID="cri-o://258c366591fe04778167b4b3595eb4316cd4f12b6070047abd9efcf1109f1470" gracePeriod=30 Dec 04 04:00:15 crc kubenswrapper[4806]: I1204 04:00:15.766570 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zrtjm" Dec 04 04:00:15 crc kubenswrapper[4806]: I1204 04:00:15.767953 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zrtjm" Dec 04 04:00:15 crc kubenswrapper[4806]: I1204 04:00:15.822523 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zrtjm" Dec 04 04:00:15 crc kubenswrapper[4806]: I1204 04:00:15.952150 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mkm6b" Dec 04 04:00:15 crc kubenswrapper[4806]: I1204 04:00:15.952484 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mkm6b" Dec 04 04:00:15 crc kubenswrapper[4806]: I1204 04:00:15.994488 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mkm6b" Dec 04 04:00:15 crc kubenswrapper[4806]: I1204 04:00:15.996114 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zrtjm" Dec 04 04:00:16 crc kubenswrapper[4806]: I1204 04:00:16.968697 4806 generic.go:334] "Generic (PLEG): container finished" podID="606e54dc-5cc4-4e02-b796-dd37b613a8d0" containerID="6be8b06358583e17b0687a963e9bb13718431db5a968e305e3eb98c49e5ce081" exitCode=0 Dec 04 04:00:16 crc kubenswrapper[4806]: I1204 04:00:16.968775 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" event={"ID":"606e54dc-5cc4-4e02-b796-dd37b613a8d0","Type":"ContainerDied","Data":"6be8b06358583e17b0687a963e9bb13718431db5a968e305e3eb98c49e5ce081"} Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.007711 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mkm6b" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.424335 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.476700 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-c67d6c44c-l7tkp"] Dec 04 04:00:17 crc kubenswrapper[4806]: E1204 04:00:17.477045 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="606e54dc-5cc4-4e02-b796-dd37b613a8d0" containerName="controller-manager" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.477060 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="606e54dc-5cc4-4e02-b796-dd37b613a8d0" containerName="controller-manager" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.477177 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="606e54dc-5cc4-4e02-b796-dd37b613a8d0" containerName="controller-manager" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.477711 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.509063 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-c67d6c44c-l7tkp"] Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.590167 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-config\") pod \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.590253 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-client-ca\") pod \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.590322 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s46kq\" (UniqueName: \"kubernetes.io/projected/606e54dc-5cc4-4e02-b796-dd37b613a8d0-kube-api-access-s46kq\") pod \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.590363 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/606e54dc-5cc4-4e02-b796-dd37b613a8d0-serving-cert\") pod \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.590416 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-proxy-ca-bundles\") pod \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\" (UID: \"606e54dc-5cc4-4e02-b796-dd37b613a8d0\") " Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.590738 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-proxy-ca-bundles\") pod \"controller-manager-c67d6c44c-l7tkp\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.590823 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-serving-cert\") pod \"controller-manager-c67d6c44c-l7tkp\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.590860 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-client-ca\") pod \"controller-manager-c67d6c44c-l7tkp\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.590942 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-config\") pod \"controller-manager-c67d6c44c-l7tkp\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.590974 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4jgb\" (UniqueName: \"kubernetes.io/projected/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-kube-api-access-t4jgb\") pod \"controller-manager-c67d6c44c-l7tkp\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.592842 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-config" (OuterVolumeSpecName: "config") pod "606e54dc-5cc4-4e02-b796-dd37b613a8d0" (UID: "606e54dc-5cc4-4e02-b796-dd37b613a8d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.593455 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-client-ca" (OuterVolumeSpecName: "client-ca") pod "606e54dc-5cc4-4e02-b796-dd37b613a8d0" (UID: "606e54dc-5cc4-4e02-b796-dd37b613a8d0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.597611 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "606e54dc-5cc4-4e02-b796-dd37b613a8d0" (UID: "606e54dc-5cc4-4e02-b796-dd37b613a8d0"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.603143 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/606e54dc-5cc4-4e02-b796-dd37b613a8d0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "606e54dc-5cc4-4e02-b796-dd37b613a8d0" (UID: "606e54dc-5cc4-4e02-b796-dd37b613a8d0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.619297 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/606e54dc-5cc4-4e02-b796-dd37b613a8d0-kube-api-access-s46kq" (OuterVolumeSpecName: "kube-api-access-s46kq") pod "606e54dc-5cc4-4e02-b796-dd37b613a8d0" (UID: "606e54dc-5cc4-4e02-b796-dd37b613a8d0"). InnerVolumeSpecName "kube-api-access-s46kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.692899 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-serving-cert\") pod \"controller-manager-c67d6c44c-l7tkp\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.692971 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-client-ca\") pod \"controller-manager-c67d6c44c-l7tkp\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.693012 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-config\") pod \"controller-manager-c67d6c44c-l7tkp\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.693038 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4jgb\" (UniqueName: \"kubernetes.io/projected/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-kube-api-access-t4jgb\") pod \"controller-manager-c67d6c44c-l7tkp\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.693081 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-proxy-ca-bundles\") pod \"controller-manager-c67d6c44c-l7tkp\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.693138 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.693151 4806 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.693161 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s46kq\" (UniqueName: \"kubernetes.io/projected/606e54dc-5cc4-4e02-b796-dd37b613a8d0-kube-api-access-s46kq\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.693173 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/606e54dc-5cc4-4e02-b796-dd37b613a8d0-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.693182 4806 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/606e54dc-5cc4-4e02-b796-dd37b613a8d0-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.694631 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-proxy-ca-bundles\") pod \"controller-manager-c67d6c44c-l7tkp\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.694880 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-config\") pod \"controller-manager-c67d6c44c-l7tkp\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.695065 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-client-ca\") pod \"controller-manager-c67d6c44c-l7tkp\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.697115 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-serving-cert\") pod \"controller-manager-c67d6c44c-l7tkp\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.714194 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4jgb\" (UniqueName: \"kubernetes.io/projected/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-kube-api-access-t4jgb\") pod \"controller-manager-c67d6c44c-l7tkp\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.794798 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.986503 4806 generic.go:334] "Generic (PLEG): container finished" podID="735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8" containerID="258c366591fe04778167b4b3595eb4316cd4f12b6070047abd9efcf1109f1470" exitCode=0 Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.987036 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" event={"ID":"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8","Type":"ContainerDied","Data":"258c366591fe04778167b4b3595eb4316cd4f12b6070047abd9efcf1109f1470"} Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.990019 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.992041 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pddk8" event={"ID":"606e54dc-5cc4-4e02-b796-dd37b613a8d0","Type":"ContainerDied","Data":"000234645f09cc051b8539795308d55d8b15a2c17f9cf79a6a11fd779da39e8b"} Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.992099 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-c67d6c44c-l7tkp"] Dec 04 04:00:17 crc kubenswrapper[4806]: I1204 04:00:17.992140 4806 scope.go:117] "RemoveContainer" containerID="6be8b06358583e17b0687a963e9bb13718431db5a968e305e3eb98c49e5ce081" Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.071036 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pddk8"] Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.091040 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pddk8"] Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.154625 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2bzpg" Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.154689 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2bzpg" Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.181714 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-c67d6c44c-l7tkp"] Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.231375 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2bzpg" Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.275255 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.402810 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-serving-cert\") pod \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\" (UID: \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\") " Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.403170 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-config\") pod \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\" (UID: \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\") " Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.403292 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-client-ca\") pod \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\" (UID: \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\") " Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.403361 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrnzn\" (UniqueName: \"kubernetes.io/projected/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-kube-api-access-zrnzn\") pod \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\" (UID: \"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8\") " Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.405022 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-config" (OuterVolumeSpecName: "config") pod "735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8" (UID: "735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.405448 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-client-ca" (OuterVolumeSpecName: "client-ca") pod "735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8" (UID: "735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.410848 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8" (UID: "735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.412950 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-kube-api-access-zrnzn" (OuterVolumeSpecName: "kube-api-access-zrnzn") pod "735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8" (UID: "735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8"). InnerVolumeSpecName "kube-api-access-zrnzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.505535 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.505607 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.505640 4806 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.505658 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrnzn\" (UniqueName: \"kubernetes.io/projected/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8-kube-api-access-zrnzn\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.588724 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kgr4l" Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.588803 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kgr4l" Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.642205 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kgr4l" Dec 04 04:00:18 crc kubenswrapper[4806]: I1204 04:00:18.998603 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" event={"ID":"735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8","Type":"ContainerDied","Data":"47649cc274c5d141b08ded1e10754ce841aec9b3694943991f738cff8e111d60"} Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:18.998669 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:18.998683 4806 scope.go:117] "RemoveContainer" containerID="258c366591fe04778167b4b3595eb4316cd4f12b6070047abd9efcf1109f1470" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.002031 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" event={"ID":"8e6b5047-e78a-46fe-9ed6-279cbc61ac52","Type":"ContainerStarted","Data":"1d27cc76b11c5fbe83a80f3fcbf6ac095b390e2c7db0d9387e046c1ccbeef6ae"} Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.002087 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" event={"ID":"8e6b5047-e78a-46fe-9ed6-279cbc61ac52","Type":"ContainerStarted","Data":"83af3de5537b6273352d4aa7dd430282d3b59e5979c760267775ec2ef2b4edda"} Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.002178 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" podUID="8e6b5047-e78a-46fe-9ed6-279cbc61ac52" containerName="controller-manager" containerID="cri-o://1d27cc76b11c5fbe83a80f3fcbf6ac095b390e2c7db0d9387e046c1ccbeef6ae" gracePeriod=30 Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.002491 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.033525 4806 patch_prober.go:28] interesting pod/controller-manager-c67d6c44c-l7tkp container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.64:8443/healthz\": read tcp 10.217.0.2:52748->10.217.0.64:8443: read: connection reset by peer" start-of-body= Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.033624 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" podUID="8e6b5047-e78a-46fe-9ed6-279cbc61ac52" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.64:8443/healthz\": read tcp 10.217.0.2:52748->10.217.0.64:8443: read: connection reset by peer" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.066776 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" podStartSLOduration=5.0667418 podStartE2EDuration="5.0667418s" podCreationTimestamp="2025-12-04 04:00:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:00:19.059614109 +0000 UTC m=+333.918127037" watchObservedRunningTime="2025-12-04 04:00:19.0667418 +0000 UTC m=+333.925254728" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.072519 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2bzpg" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.080373 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kgr4l" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.083411 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml"] Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.088315 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-f68ml"] Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.432717 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="606e54dc-5cc4-4e02-b796-dd37b613a8d0" path="/var/lib/kubelet/pods/606e54dc-5cc4-4e02-b796-dd37b613a8d0/volumes" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.433612 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8" path="/var/lib/kubelet/pods/735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8/volumes" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.516140 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz"] Dec 04 04:00:19 crc kubenswrapper[4806]: E1204 04:00:19.516460 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8" containerName="route-controller-manager" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.516482 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8" containerName="route-controller-manager" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.516615 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="735f26b7-463e-4ae2-94d4-e7bd0b0ac2b8" containerName="route-controller-manager" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.517206 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.520103 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.520297 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.521262 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.522364 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.522402 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.522578 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.533533 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz"] Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.647467 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58f49065-a62e-4fd9-928c-125ac7f1a83a-serving-cert\") pod \"route-controller-manager-749699db6b-cf8hz\" (UID: \"58f49065-a62e-4fd9-928c-125ac7f1a83a\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.647965 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58f49065-a62e-4fd9-928c-125ac7f1a83a-config\") pod \"route-controller-manager-749699db6b-cf8hz\" (UID: \"58f49065-a62e-4fd9-928c-125ac7f1a83a\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.648147 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58f49065-a62e-4fd9-928c-125ac7f1a83a-client-ca\") pod \"route-controller-manager-749699db6b-cf8hz\" (UID: \"58f49065-a62e-4fd9-928c-125ac7f1a83a\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.648265 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7hkq\" (UniqueName: \"kubernetes.io/projected/58f49065-a62e-4fd9-928c-125ac7f1a83a-kube-api-access-r7hkq\") pod \"route-controller-manager-749699db6b-cf8hz\" (UID: \"58f49065-a62e-4fd9-928c-125ac7f1a83a\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.750095 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58f49065-a62e-4fd9-928c-125ac7f1a83a-client-ca\") pod \"route-controller-manager-749699db6b-cf8hz\" (UID: \"58f49065-a62e-4fd9-928c-125ac7f1a83a\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.750665 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7hkq\" (UniqueName: \"kubernetes.io/projected/58f49065-a62e-4fd9-928c-125ac7f1a83a-kube-api-access-r7hkq\") pod \"route-controller-manager-749699db6b-cf8hz\" (UID: \"58f49065-a62e-4fd9-928c-125ac7f1a83a\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.750713 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58f49065-a62e-4fd9-928c-125ac7f1a83a-serving-cert\") pod \"route-controller-manager-749699db6b-cf8hz\" (UID: \"58f49065-a62e-4fd9-928c-125ac7f1a83a\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.750754 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58f49065-a62e-4fd9-928c-125ac7f1a83a-config\") pod \"route-controller-manager-749699db6b-cf8hz\" (UID: \"58f49065-a62e-4fd9-928c-125ac7f1a83a\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.752180 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58f49065-a62e-4fd9-928c-125ac7f1a83a-client-ca\") pod \"route-controller-manager-749699db6b-cf8hz\" (UID: \"58f49065-a62e-4fd9-928c-125ac7f1a83a\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.752251 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58f49065-a62e-4fd9-928c-125ac7f1a83a-config\") pod \"route-controller-manager-749699db6b-cf8hz\" (UID: \"58f49065-a62e-4fd9-928c-125ac7f1a83a\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.759993 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58f49065-a62e-4fd9-928c-125ac7f1a83a-serving-cert\") pod \"route-controller-manager-749699db6b-cf8hz\" (UID: \"58f49065-a62e-4fd9-928c-125ac7f1a83a\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.768529 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7hkq\" (UniqueName: \"kubernetes.io/projected/58f49065-a62e-4fd9-928c-125ac7f1a83a-kube-api-access-r7hkq\") pod \"route-controller-manager-749699db6b-cf8hz\" (UID: \"58f49065-a62e-4fd9-928c-125ac7f1a83a\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.833856 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:19 crc kubenswrapper[4806]: I1204 04:00:19.964379 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz"] Dec 04 04:00:20 crc kubenswrapper[4806]: I1204 04:00:20.080096 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz"] Dec 04 04:00:20 crc kubenswrapper[4806]: W1204 04:00:20.087314 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58f49065_a62e_4fd9_928c_125ac7f1a83a.slice/crio-26be80c215d561b938b61db02f86d6fd8d567fc5713d4eb047b796c1f342b40c WatchSource:0}: Error finding container 26be80c215d561b938b61db02f86d6fd8d567fc5713d4eb047b796c1f342b40c: Status 404 returned error can't find the container with id 26be80c215d561b938b61db02f86d6fd8d567fc5713d4eb047b796c1f342b40c Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.023030 4806 generic.go:334] "Generic (PLEG): container finished" podID="8e6b5047-e78a-46fe-9ed6-279cbc61ac52" containerID="1d27cc76b11c5fbe83a80f3fcbf6ac095b390e2c7db0d9387e046c1ccbeef6ae" exitCode=0 Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.023118 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" event={"ID":"8e6b5047-e78a-46fe-9ed6-279cbc61ac52","Type":"ContainerDied","Data":"1d27cc76b11c5fbe83a80f3fcbf6ac095b390e2c7db0d9387e046c1ccbeef6ae"} Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.026756 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" event={"ID":"58f49065-a62e-4fd9-928c-125ac7f1a83a","Type":"ContainerStarted","Data":"bba5ea51118ddf627adfe0ad4f640d71eb30375f7507641c50cd130a87e602e8"} Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.026822 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" event={"ID":"58f49065-a62e-4fd9-928c-125ac7f1a83a","Type":"ContainerStarted","Data":"26be80c215d561b938b61db02f86d6fd8d567fc5713d4eb047b796c1f342b40c"} Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.027005 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" podUID="58f49065-a62e-4fd9-928c-125ac7f1a83a" containerName="route-controller-manager" containerID="cri-o://bba5ea51118ddf627adfe0ad4f640d71eb30375f7507641c50cd130a87e602e8" gracePeriod=30 Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.027212 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.068112 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" podStartSLOduration=3.068086315 podStartE2EDuration="3.068086315s" podCreationTimestamp="2025-12-04 04:00:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:00:21.067846487 +0000 UTC m=+335.926359415" watchObservedRunningTime="2025-12-04 04:00:21.068086315 +0000 UTC m=+335.926599243" Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.179572 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.274674 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-client-ca\") pod \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.274851 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4jgb\" (UniqueName: \"kubernetes.io/projected/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-kube-api-access-t4jgb\") pod \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.274911 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-serving-cert\") pod \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.274985 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-config\") pod \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.275019 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-proxy-ca-bundles\") pod \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\" (UID: \"8e6b5047-e78a-46fe-9ed6-279cbc61ac52\") " Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.275882 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-client-ca" (OuterVolumeSpecName: "client-ca") pod "8e6b5047-e78a-46fe-9ed6-279cbc61ac52" (UID: "8e6b5047-e78a-46fe-9ed6-279cbc61ac52"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.276083 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8e6b5047-e78a-46fe-9ed6-279cbc61ac52" (UID: "8e6b5047-e78a-46fe-9ed6-279cbc61ac52"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.276420 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-config" (OuterVolumeSpecName: "config") pod "8e6b5047-e78a-46fe-9ed6-279cbc61ac52" (UID: "8e6b5047-e78a-46fe-9ed6-279cbc61ac52"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.297352 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-kube-api-access-t4jgb" (OuterVolumeSpecName: "kube-api-access-t4jgb") pod "8e6b5047-e78a-46fe-9ed6-279cbc61ac52" (UID: "8e6b5047-e78a-46fe-9ed6-279cbc61ac52"). InnerVolumeSpecName "kube-api-access-t4jgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.299178 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8e6b5047-e78a-46fe-9ed6-279cbc61ac52" (UID: "8e6b5047-e78a-46fe-9ed6-279cbc61ac52"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.377471 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4jgb\" (UniqueName: \"kubernetes.io/projected/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-kube-api-access-t4jgb\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.377523 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.377544 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.377560 4806 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.377573 4806 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8e6b5047-e78a-46fe-9ed6-279cbc61ac52-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.419499 4806 patch_prober.go:28] interesting pod/route-controller-manager-749699db6b-cf8hz container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": read tcp 10.217.0.2:44246->10.217.0.65:8443: read: connection reset by peer" start-of-body= Dec 04 04:00:21 crc kubenswrapper[4806]: I1204 04:00:21.419592 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" podUID="58f49065-a62e-4fd9-928c-125ac7f1a83a" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": read tcp 10.217.0.2:44246->10.217.0.65:8443: read: connection reset by peer" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.047507 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-749699db6b-cf8hz_58f49065-a62e-4fd9-928c-125ac7f1a83a/route-controller-manager/0.log" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.047571 4806 generic.go:334] "Generic (PLEG): container finished" podID="58f49065-a62e-4fd9-928c-125ac7f1a83a" containerID="bba5ea51118ddf627adfe0ad4f640d71eb30375f7507641c50cd130a87e602e8" exitCode=255 Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.047655 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" event={"ID":"58f49065-a62e-4fd9-928c-125ac7f1a83a","Type":"ContainerDied","Data":"bba5ea51118ddf627adfe0ad4f640d71eb30375f7507641c50cd130a87e602e8"} Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.049425 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" event={"ID":"8e6b5047-e78a-46fe-9ed6-279cbc61ac52","Type":"ContainerDied","Data":"83af3de5537b6273352d4aa7dd430282d3b59e5979c760267775ec2ef2b4edda"} Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.049466 4806 scope.go:117] "RemoveContainer" containerID="1d27cc76b11c5fbe83a80f3fcbf6ac095b390e2c7db0d9387e046c1ccbeef6ae" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.049618 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c67d6c44c-l7tkp" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.080013 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-c67d6c44c-l7tkp"] Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.082839 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-c67d6c44c-l7tkp"] Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.171104 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-749699db6b-cf8hz_58f49065-a62e-4fd9-928c-125ac7f1a83a/route-controller-manager/0.log" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.171196 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.199690 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58f49065-a62e-4fd9-928c-125ac7f1a83a-config\") pod \"58f49065-a62e-4fd9-928c-125ac7f1a83a\" (UID: \"58f49065-a62e-4fd9-928c-125ac7f1a83a\") " Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.199823 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58f49065-a62e-4fd9-928c-125ac7f1a83a-client-ca\") pod \"58f49065-a62e-4fd9-928c-125ac7f1a83a\" (UID: \"58f49065-a62e-4fd9-928c-125ac7f1a83a\") " Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.199857 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7hkq\" (UniqueName: \"kubernetes.io/projected/58f49065-a62e-4fd9-928c-125ac7f1a83a-kube-api-access-r7hkq\") pod \"58f49065-a62e-4fd9-928c-125ac7f1a83a\" (UID: \"58f49065-a62e-4fd9-928c-125ac7f1a83a\") " Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.199889 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58f49065-a62e-4fd9-928c-125ac7f1a83a-serving-cert\") pod \"58f49065-a62e-4fd9-928c-125ac7f1a83a\" (UID: \"58f49065-a62e-4fd9-928c-125ac7f1a83a\") " Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.200911 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58f49065-a62e-4fd9-928c-125ac7f1a83a-config" (OuterVolumeSpecName: "config") pod "58f49065-a62e-4fd9-928c-125ac7f1a83a" (UID: "58f49065-a62e-4fd9-928c-125ac7f1a83a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.200961 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58f49065-a62e-4fd9-928c-125ac7f1a83a-client-ca" (OuterVolumeSpecName: "client-ca") pod "58f49065-a62e-4fd9-928c-125ac7f1a83a" (UID: "58f49065-a62e-4fd9-928c-125ac7f1a83a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.204470 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58f49065-a62e-4fd9-928c-125ac7f1a83a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "58f49065-a62e-4fd9-928c-125ac7f1a83a" (UID: "58f49065-a62e-4fd9-928c-125ac7f1a83a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.208863 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58f49065-a62e-4fd9-928c-125ac7f1a83a-kube-api-access-r7hkq" (OuterVolumeSpecName: "kube-api-access-r7hkq") pod "58f49065-a62e-4fd9-928c-125ac7f1a83a" (UID: "58f49065-a62e-4fd9-928c-125ac7f1a83a"). InnerVolumeSpecName "kube-api-access-r7hkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.301721 4806 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58f49065-a62e-4fd9-928c-125ac7f1a83a-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.301772 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7hkq\" (UniqueName: \"kubernetes.io/projected/58f49065-a62e-4fd9-928c-125ac7f1a83a-kube-api-access-r7hkq\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.301786 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58f49065-a62e-4fd9-928c-125ac7f1a83a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.301796 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58f49065-a62e-4fd9-928c-125ac7f1a83a-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.518223 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7798c9bf8-7rbp2"] Dec 04 04:00:22 crc kubenswrapper[4806]: E1204 04:00:22.518708 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58f49065-a62e-4fd9-928c-125ac7f1a83a" containerName="route-controller-manager" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.518733 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="58f49065-a62e-4fd9-928c-125ac7f1a83a" containerName="route-controller-manager" Dec 04 04:00:22 crc kubenswrapper[4806]: E1204 04:00:22.518747 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6b5047-e78a-46fe-9ed6-279cbc61ac52" containerName="controller-manager" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.518756 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6b5047-e78a-46fe-9ed6-279cbc61ac52" containerName="controller-manager" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.518972 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e6b5047-e78a-46fe-9ed6-279cbc61ac52" containerName="controller-manager" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.518986 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="58f49065-a62e-4fd9-928c-125ac7f1a83a" containerName="route-controller-manager" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.519583 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.521815 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.522044 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.524834 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.525205 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.529812 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.529823 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.530094 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.535951 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7798c9bf8-7rbp2"] Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.606396 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-config\") pod \"controller-manager-7798c9bf8-7rbp2\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.606520 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wwps\" (UniqueName: \"kubernetes.io/projected/6ddca849-0244-44c9-9917-bf3a28c3d477-kube-api-access-5wwps\") pod \"controller-manager-7798c9bf8-7rbp2\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.606585 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-proxy-ca-bundles\") pod \"controller-manager-7798c9bf8-7rbp2\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.606618 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-client-ca\") pod \"controller-manager-7798c9bf8-7rbp2\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.606645 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ddca849-0244-44c9-9917-bf3a28c3d477-serving-cert\") pod \"controller-manager-7798c9bf8-7rbp2\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.707690 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ddca849-0244-44c9-9917-bf3a28c3d477-serving-cert\") pod \"controller-manager-7798c9bf8-7rbp2\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.708270 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-config\") pod \"controller-manager-7798c9bf8-7rbp2\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.708321 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wwps\" (UniqueName: \"kubernetes.io/projected/6ddca849-0244-44c9-9917-bf3a28c3d477-kube-api-access-5wwps\") pod \"controller-manager-7798c9bf8-7rbp2\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.708365 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-proxy-ca-bundles\") pod \"controller-manager-7798c9bf8-7rbp2\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.708384 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-client-ca\") pod \"controller-manager-7798c9bf8-7rbp2\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.709593 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-client-ca\") pod \"controller-manager-7798c9bf8-7rbp2\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.709816 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-config\") pod \"controller-manager-7798c9bf8-7rbp2\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.710394 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-proxy-ca-bundles\") pod \"controller-manager-7798c9bf8-7rbp2\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.713233 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ddca849-0244-44c9-9917-bf3a28c3d477-serving-cert\") pod \"controller-manager-7798c9bf8-7rbp2\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.728747 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wwps\" (UniqueName: \"kubernetes.io/projected/6ddca849-0244-44c9-9917-bf3a28c3d477-kube-api-access-5wwps\") pod \"controller-manager-7798c9bf8-7rbp2\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:22 crc kubenswrapper[4806]: I1204 04:00:22.894661 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.070905 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-749699db6b-cf8hz_58f49065-a62e-4fd9-928c-125ac7f1a83a/route-controller-manager/0.log" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.071003 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" event={"ID":"58f49065-a62e-4fd9-928c-125ac7f1a83a","Type":"ContainerDied","Data":"26be80c215d561b938b61db02f86d6fd8d567fc5713d4eb047b796c1f342b40c"} Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.071072 4806 scope.go:117] "RemoveContainer" containerID="bba5ea51118ddf627adfe0ad4f640d71eb30375f7507641c50cd130a87e602e8" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.071070 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.114408 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7798c9bf8-7rbp2"] Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.121112 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz"] Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.128715 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-749699db6b-cf8hz"] Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.433805 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58f49065-a62e-4fd9-928c-125ac7f1a83a" path="/var/lib/kubelet/pods/58f49065-a62e-4fd9-928c-125ac7f1a83a/volumes" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.435104 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e6b5047-e78a-46fe-9ed6-279cbc61ac52" path="/var/lib/kubelet/pods/8e6b5047-e78a-46fe-9ed6-279cbc61ac52/volumes" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.520785 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4"] Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.522037 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.527518 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.527781 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.530907 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.532370 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.541880 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.542388 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.560367 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4"] Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.723355 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/48e63397-f90d-4325-b5ce-4e96dcef6c19-client-ca\") pod \"route-controller-manager-795597c8b4-m75l4\" (UID: \"48e63397-f90d-4325-b5ce-4e96dcef6c19\") " pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.723517 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48e63397-f90d-4325-b5ce-4e96dcef6c19-serving-cert\") pod \"route-controller-manager-795597c8b4-m75l4\" (UID: \"48e63397-f90d-4325-b5ce-4e96dcef6c19\") " pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.723573 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48e63397-f90d-4325-b5ce-4e96dcef6c19-config\") pod \"route-controller-manager-795597c8b4-m75l4\" (UID: \"48e63397-f90d-4325-b5ce-4e96dcef6c19\") " pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.723636 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btlqj\" (UniqueName: \"kubernetes.io/projected/48e63397-f90d-4325-b5ce-4e96dcef6c19-kube-api-access-btlqj\") pod \"route-controller-manager-795597c8b4-m75l4\" (UID: \"48e63397-f90d-4325-b5ce-4e96dcef6c19\") " pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.824818 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btlqj\" (UniqueName: \"kubernetes.io/projected/48e63397-f90d-4325-b5ce-4e96dcef6c19-kube-api-access-btlqj\") pod \"route-controller-manager-795597c8b4-m75l4\" (UID: \"48e63397-f90d-4325-b5ce-4e96dcef6c19\") " pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.824905 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/48e63397-f90d-4325-b5ce-4e96dcef6c19-client-ca\") pod \"route-controller-manager-795597c8b4-m75l4\" (UID: \"48e63397-f90d-4325-b5ce-4e96dcef6c19\") " pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.825002 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48e63397-f90d-4325-b5ce-4e96dcef6c19-serving-cert\") pod \"route-controller-manager-795597c8b4-m75l4\" (UID: \"48e63397-f90d-4325-b5ce-4e96dcef6c19\") " pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.825033 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48e63397-f90d-4325-b5ce-4e96dcef6c19-config\") pod \"route-controller-manager-795597c8b4-m75l4\" (UID: \"48e63397-f90d-4325-b5ce-4e96dcef6c19\") " pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.825935 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/48e63397-f90d-4325-b5ce-4e96dcef6c19-client-ca\") pod \"route-controller-manager-795597c8b4-m75l4\" (UID: \"48e63397-f90d-4325-b5ce-4e96dcef6c19\") " pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.826295 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48e63397-f90d-4325-b5ce-4e96dcef6c19-config\") pod \"route-controller-manager-795597c8b4-m75l4\" (UID: \"48e63397-f90d-4325-b5ce-4e96dcef6c19\") " pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.840638 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48e63397-f90d-4325-b5ce-4e96dcef6c19-serving-cert\") pod \"route-controller-manager-795597c8b4-m75l4\" (UID: \"48e63397-f90d-4325-b5ce-4e96dcef6c19\") " pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:00:23 crc kubenswrapper[4806]: I1204 04:00:23.856730 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btlqj\" (UniqueName: \"kubernetes.io/projected/48e63397-f90d-4325-b5ce-4e96dcef6c19-kube-api-access-btlqj\") pod \"route-controller-manager-795597c8b4-m75l4\" (UID: \"48e63397-f90d-4325-b5ce-4e96dcef6c19\") " pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:00:24 crc kubenswrapper[4806]: I1204 04:00:24.079074 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" event={"ID":"6ddca849-0244-44c9-9917-bf3a28c3d477","Type":"ContainerStarted","Data":"bfb42a576a40b0d1c0a9e0d7fcf9a6c59a3320d5c63147a1f733059c12fcfcd4"} Dec 04 04:00:24 crc kubenswrapper[4806]: I1204 04:00:24.079140 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" event={"ID":"6ddca849-0244-44c9-9917-bf3a28c3d477","Type":"ContainerStarted","Data":"0b22ba01fdbd6209b85024cdcc0eee7dea1f4271a0ddd9b7e46fe77c4e4e8fc1"} Dec 04 04:00:24 crc kubenswrapper[4806]: I1204 04:00:24.080401 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:24 crc kubenswrapper[4806]: I1204 04:00:24.086963 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:24 crc kubenswrapper[4806]: I1204 04:00:24.105807 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" podStartSLOduration=6.105777421 podStartE2EDuration="6.105777421s" podCreationTimestamp="2025-12-04 04:00:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:00:24.099126805 +0000 UTC m=+338.957639733" watchObservedRunningTime="2025-12-04 04:00:24.105777421 +0000 UTC m=+338.964290349" Dec 04 04:00:24 crc kubenswrapper[4806]: I1204 04:00:24.143195 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:00:24 crc kubenswrapper[4806]: I1204 04:00:24.441159 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4"] Dec 04 04:00:25 crc kubenswrapper[4806]: I1204 04:00:25.088054 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" event={"ID":"48e63397-f90d-4325-b5ce-4e96dcef6c19","Type":"ContainerStarted","Data":"20ece38be669e0adb340a882012a0299d4fc0a0106e4106917fa07580ba6783b"} Dec 04 04:00:25 crc kubenswrapper[4806]: I1204 04:00:25.088480 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" event={"ID":"48e63397-f90d-4325-b5ce-4e96dcef6c19","Type":"ContainerStarted","Data":"24c6fc1a15a5d7174b521f35cc341a2e0e5f3f72e2266a541e7c5dc13f1187c7"} Dec 04 04:00:25 crc kubenswrapper[4806]: I1204 04:00:25.088740 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:00:25 crc kubenswrapper[4806]: I1204 04:00:25.112111 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" podStartSLOduration=6.112077516 podStartE2EDuration="6.112077516s" podCreationTimestamp="2025-12-04 04:00:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:00:25.109748861 +0000 UTC m=+339.968261789" watchObservedRunningTime="2025-12-04 04:00:25.112077516 +0000 UTC m=+339.970590444" Dec 04 04:00:25 crc kubenswrapper[4806]: I1204 04:00:25.285754 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:00:27 crc kubenswrapper[4806]: I1204 04:00:27.047396 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:00:27 crc kubenswrapper[4806]: I1204 04:00:27.047550 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:00:54 crc kubenswrapper[4806]: I1204 04:00:54.917423 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7798c9bf8-7rbp2"] Dec 04 04:00:54 crc kubenswrapper[4806]: I1204 04:00:54.918590 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" podUID="6ddca849-0244-44c9-9917-bf3a28c3d477" containerName="controller-manager" containerID="cri-o://bfb42a576a40b0d1c0a9e0d7fcf9a6c59a3320d5c63147a1f733059c12fcfcd4" gracePeriod=30 Dec 04 04:00:55 crc kubenswrapper[4806]: I1204 04:00:55.285660 4806 generic.go:334] "Generic (PLEG): container finished" podID="6ddca849-0244-44c9-9917-bf3a28c3d477" containerID="bfb42a576a40b0d1c0a9e0d7fcf9a6c59a3320d5c63147a1f733059c12fcfcd4" exitCode=0 Dec 04 04:00:55 crc kubenswrapper[4806]: I1204 04:00:55.285847 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" event={"ID":"6ddca849-0244-44c9-9917-bf3a28c3d477","Type":"ContainerDied","Data":"bfb42a576a40b0d1c0a9e0d7fcf9a6c59a3320d5c63147a1f733059c12fcfcd4"} Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.294519 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" event={"ID":"6ddca849-0244-44c9-9917-bf3a28c3d477","Type":"ContainerDied","Data":"0b22ba01fdbd6209b85024cdcc0eee7dea1f4271a0ddd9b7e46fe77c4e4e8fc1"} Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.295426 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b22ba01fdbd6209b85024cdcc0eee7dea1f4271a0ddd9b7e46fe77c4e4e8fc1" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.590426 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.673107 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-64985d5fdb-nzv9x"] Dec 04 04:00:56 crc kubenswrapper[4806]: E1204 04:00:56.673728 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ddca849-0244-44c9-9917-bf3a28c3d477" containerName="controller-manager" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.673755 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ddca849-0244-44c9-9917-bf3a28c3d477" containerName="controller-manager" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.693428 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ddca849-0244-44c9-9917-bf3a28c3d477" containerName="controller-manager" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.694376 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.719827 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-64985d5fdb-nzv9x"] Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.774573 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-config\") pod \"6ddca849-0244-44c9-9917-bf3a28c3d477\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.774678 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-proxy-ca-bundles\") pod \"6ddca849-0244-44c9-9917-bf3a28c3d477\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.774716 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ddca849-0244-44c9-9917-bf3a28c3d477-serving-cert\") pod \"6ddca849-0244-44c9-9917-bf3a28c3d477\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.774846 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-client-ca\") pod \"6ddca849-0244-44c9-9917-bf3a28c3d477\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.774875 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wwps\" (UniqueName: \"kubernetes.io/projected/6ddca849-0244-44c9-9917-bf3a28c3d477-kube-api-access-5wwps\") pod \"6ddca849-0244-44c9-9917-bf3a28c3d477\" (UID: \"6ddca849-0244-44c9-9917-bf3a28c3d477\") " Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.777702 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-config" (OuterVolumeSpecName: "config") pod "6ddca849-0244-44c9-9917-bf3a28c3d477" (UID: "6ddca849-0244-44c9-9917-bf3a28c3d477"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.777812 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "6ddca849-0244-44c9-9917-bf3a28c3d477" (UID: "6ddca849-0244-44c9-9917-bf3a28c3d477"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.778239 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-client-ca" (OuterVolumeSpecName: "client-ca") pod "6ddca849-0244-44c9-9917-bf3a28c3d477" (UID: "6ddca849-0244-44c9-9917-bf3a28c3d477"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.790659 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ddca849-0244-44c9-9917-bf3a28c3d477-kube-api-access-5wwps" (OuterVolumeSpecName: "kube-api-access-5wwps") pod "6ddca849-0244-44c9-9917-bf3a28c3d477" (UID: "6ddca849-0244-44c9-9917-bf3a28c3d477"). InnerVolumeSpecName "kube-api-access-5wwps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.801185 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ddca849-0244-44c9-9917-bf3a28c3d477-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6ddca849-0244-44c9-9917-bf3a28c3d477" (UID: "6ddca849-0244-44c9-9917-bf3a28c3d477"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.876453 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b662aa82-250d-4f36-ae39-ede483472442-serving-cert\") pod \"controller-manager-64985d5fdb-nzv9x\" (UID: \"b662aa82-250d-4f36-ae39-ede483472442\") " pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.876554 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b662aa82-250d-4f36-ae39-ede483472442-proxy-ca-bundles\") pod \"controller-manager-64985d5fdb-nzv9x\" (UID: \"b662aa82-250d-4f36-ae39-ede483472442\") " pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.876767 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b662aa82-250d-4f36-ae39-ede483472442-client-ca\") pod \"controller-manager-64985d5fdb-nzv9x\" (UID: \"b662aa82-250d-4f36-ae39-ede483472442\") " pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.876944 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5kh4\" (UniqueName: \"kubernetes.io/projected/b662aa82-250d-4f36-ae39-ede483472442-kube-api-access-w5kh4\") pod \"controller-manager-64985d5fdb-nzv9x\" (UID: \"b662aa82-250d-4f36-ae39-ede483472442\") " pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.877191 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b662aa82-250d-4f36-ae39-ede483472442-config\") pod \"controller-manager-64985d5fdb-nzv9x\" (UID: \"b662aa82-250d-4f36-ae39-ede483472442\") " pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.877382 4806 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.877409 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ddca849-0244-44c9-9917-bf3a28c3d477-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.877423 4806 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.877436 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wwps\" (UniqueName: \"kubernetes.io/projected/6ddca849-0244-44c9-9917-bf3a28c3d477-kube-api-access-5wwps\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.877454 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ddca849-0244-44c9-9917-bf3a28c3d477-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.979309 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b662aa82-250d-4f36-ae39-ede483472442-serving-cert\") pod \"controller-manager-64985d5fdb-nzv9x\" (UID: \"b662aa82-250d-4f36-ae39-ede483472442\") " pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.979393 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b662aa82-250d-4f36-ae39-ede483472442-proxy-ca-bundles\") pod \"controller-manager-64985d5fdb-nzv9x\" (UID: \"b662aa82-250d-4f36-ae39-ede483472442\") " pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.979429 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b662aa82-250d-4f36-ae39-ede483472442-client-ca\") pod \"controller-manager-64985d5fdb-nzv9x\" (UID: \"b662aa82-250d-4f36-ae39-ede483472442\") " pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.979461 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5kh4\" (UniqueName: \"kubernetes.io/projected/b662aa82-250d-4f36-ae39-ede483472442-kube-api-access-w5kh4\") pod \"controller-manager-64985d5fdb-nzv9x\" (UID: \"b662aa82-250d-4f36-ae39-ede483472442\") " pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.979521 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b662aa82-250d-4f36-ae39-ede483472442-config\") pod \"controller-manager-64985d5fdb-nzv9x\" (UID: \"b662aa82-250d-4f36-ae39-ede483472442\") " pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.981007 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b662aa82-250d-4f36-ae39-ede483472442-proxy-ca-bundles\") pod \"controller-manager-64985d5fdb-nzv9x\" (UID: \"b662aa82-250d-4f36-ae39-ede483472442\") " pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.981031 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b662aa82-250d-4f36-ae39-ede483472442-client-ca\") pod \"controller-manager-64985d5fdb-nzv9x\" (UID: \"b662aa82-250d-4f36-ae39-ede483472442\") " pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.981527 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b662aa82-250d-4f36-ae39-ede483472442-config\") pod \"controller-manager-64985d5fdb-nzv9x\" (UID: \"b662aa82-250d-4f36-ae39-ede483472442\") " pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.985999 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b662aa82-250d-4f36-ae39-ede483472442-serving-cert\") pod \"controller-manager-64985d5fdb-nzv9x\" (UID: \"b662aa82-250d-4f36-ae39-ede483472442\") " pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:56 crc kubenswrapper[4806]: I1204 04:00:56.999376 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5kh4\" (UniqueName: \"kubernetes.io/projected/b662aa82-250d-4f36-ae39-ede483472442-kube-api-access-w5kh4\") pod \"controller-manager-64985d5fdb-nzv9x\" (UID: \"b662aa82-250d-4f36-ae39-ede483472442\") " pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:57 crc kubenswrapper[4806]: I1204 04:00:57.047022 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:00:57 crc kubenswrapper[4806]: I1204 04:00:57.047094 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:00:57 crc kubenswrapper[4806]: I1204 04:00:57.072308 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:57 crc kubenswrapper[4806]: I1204 04:00:57.307131 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7798c9bf8-7rbp2" Dec 04 04:00:57 crc kubenswrapper[4806]: I1204 04:00:57.345858 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-64985d5fdb-nzv9x"] Dec 04 04:00:57 crc kubenswrapper[4806]: I1204 04:00:57.380999 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7798c9bf8-7rbp2"] Dec 04 04:00:57 crc kubenswrapper[4806]: I1204 04:00:57.387066 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7798c9bf8-7rbp2"] Dec 04 04:00:57 crc kubenswrapper[4806]: I1204 04:00:57.438437 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ddca849-0244-44c9-9917-bf3a28c3d477" path="/var/lib/kubelet/pods/6ddca849-0244-44c9-9917-bf3a28c3d477/volumes" Dec 04 04:00:58 crc kubenswrapper[4806]: I1204 04:00:58.316352 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" event={"ID":"b662aa82-250d-4f36-ae39-ede483472442","Type":"ContainerStarted","Data":"8a91c02a9b1d1afd806c0e4180f6219e147143c79f91610e3d5976fc5c392435"} Dec 04 04:00:58 crc kubenswrapper[4806]: I1204 04:00:58.316977 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" event={"ID":"b662aa82-250d-4f36-ae39-ede483472442","Type":"ContainerStarted","Data":"40768dbc8c0473932903b618650bbd21b01400935176b60c0e5f7dd527ce7343"} Dec 04 04:00:58 crc kubenswrapper[4806]: I1204 04:00:58.317010 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:58 crc kubenswrapper[4806]: I1204 04:00:58.322035 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" Dec 04 04:00:58 crc kubenswrapper[4806]: I1204 04:00:58.338976 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-64985d5fdb-nzv9x" podStartSLOduration=4.3389458770000005 podStartE2EDuration="4.338945877s" podCreationTimestamp="2025-12-04 04:00:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:00:58.337961375 +0000 UTC m=+373.196474303" watchObservedRunningTime="2025-12-04 04:00:58.338945877 +0000 UTC m=+373.197458835" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.552465 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9vsks"] Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.553911 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.578663 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9vsks"] Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.681227 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1ebbaddc-923d-4d2c-b97f-bba625839700-registry-tls\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.681301 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zh6k\" (UniqueName: \"kubernetes.io/projected/1ebbaddc-923d-4d2c-b97f-bba625839700-kube-api-access-4zh6k\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.681332 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ebbaddc-923d-4d2c-b97f-bba625839700-trusted-ca\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.681351 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1ebbaddc-923d-4d2c-b97f-bba625839700-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.681385 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.681405 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1ebbaddc-923d-4d2c-b97f-bba625839700-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.681430 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1ebbaddc-923d-4d2c-b97f-bba625839700-registry-certificates\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.681473 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1ebbaddc-923d-4d2c-b97f-bba625839700-bound-sa-token\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.749711 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.783231 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1ebbaddc-923d-4d2c-b97f-bba625839700-registry-tls\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.783329 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zh6k\" (UniqueName: \"kubernetes.io/projected/1ebbaddc-923d-4d2c-b97f-bba625839700-kube-api-access-4zh6k\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.783379 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ebbaddc-923d-4d2c-b97f-bba625839700-trusted-ca\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.783408 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1ebbaddc-923d-4d2c-b97f-bba625839700-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.783448 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1ebbaddc-923d-4d2c-b97f-bba625839700-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.783480 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1ebbaddc-923d-4d2c-b97f-bba625839700-registry-certificates\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.783508 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1ebbaddc-923d-4d2c-b97f-bba625839700-bound-sa-token\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.786595 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1ebbaddc-923d-4d2c-b97f-bba625839700-trusted-ca\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.786902 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1ebbaddc-923d-4d2c-b97f-bba625839700-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.787424 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1ebbaddc-923d-4d2c-b97f-bba625839700-registry-certificates\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.792311 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1ebbaddc-923d-4d2c-b97f-bba625839700-registry-tls\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.793341 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1ebbaddc-923d-4d2c-b97f-bba625839700-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.817912 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zh6k\" (UniqueName: \"kubernetes.io/projected/1ebbaddc-923d-4d2c-b97f-bba625839700-kube-api-access-4zh6k\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.818884 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1ebbaddc-923d-4d2c-b97f-bba625839700-bound-sa-token\") pod \"image-registry-66df7c8f76-9vsks\" (UID: \"1ebbaddc-923d-4d2c-b97f-bba625839700\") " pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:02 crc kubenswrapper[4806]: I1204 04:01:02.871391 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:03 crc kubenswrapper[4806]: I1204 04:01:03.343220 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9vsks"] Dec 04 04:01:04 crc kubenswrapper[4806]: I1204 04:01:04.356047 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" event={"ID":"1ebbaddc-923d-4d2c-b97f-bba625839700","Type":"ContainerStarted","Data":"1ff2d1703e42fe6bcf0eb059bd8a080775928a6e1c44ba577e9450e498fc5b70"} Dec 04 04:01:04 crc kubenswrapper[4806]: I1204 04:01:04.356567 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:04 crc kubenswrapper[4806]: I1204 04:01:04.356581 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" event={"ID":"1ebbaddc-923d-4d2c-b97f-bba625839700","Type":"ContainerStarted","Data":"0710c0021d64d250a01f92f9d5df694d26f926d541ba2cede273d8452bd482d1"} Dec 04 04:01:04 crc kubenswrapper[4806]: I1204 04:01:04.380892 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" podStartSLOduration=2.380865261 podStartE2EDuration="2.380865261s" podCreationTimestamp="2025-12-04 04:01:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:01:04.375446265 +0000 UTC m=+379.233959193" watchObservedRunningTime="2025-12-04 04:01:04.380865261 +0000 UTC m=+379.239378189" Dec 04 04:01:14 crc kubenswrapper[4806]: I1204 04:01:14.939597 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4"] Dec 04 04:01:14 crc kubenswrapper[4806]: I1204 04:01:14.940907 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" podUID="48e63397-f90d-4325-b5ce-4e96dcef6c19" containerName="route-controller-manager" containerID="cri-o://20ece38be669e0adb340a882012a0299d4fc0a0106e4106917fa07580ba6783b" gracePeriod=30 Dec 04 04:01:15 crc kubenswrapper[4806]: I1204 04:01:15.425415 4806 generic.go:334] "Generic (PLEG): container finished" podID="48e63397-f90d-4325-b5ce-4e96dcef6c19" containerID="20ece38be669e0adb340a882012a0299d4fc0a0106e4106917fa07580ba6783b" exitCode=0 Dec 04 04:01:15 crc kubenswrapper[4806]: I1204 04:01:15.430378 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" event={"ID":"48e63397-f90d-4325-b5ce-4e96dcef6c19","Type":"ContainerDied","Data":"20ece38be669e0adb340a882012a0299d4fc0a0106e4106917fa07580ba6783b"} Dec 04 04:01:15 crc kubenswrapper[4806]: I1204 04:01:15.896712 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.000233 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48e63397-f90d-4325-b5ce-4e96dcef6c19-config\") pod \"48e63397-f90d-4325-b5ce-4e96dcef6c19\" (UID: \"48e63397-f90d-4325-b5ce-4e96dcef6c19\") " Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.000444 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48e63397-f90d-4325-b5ce-4e96dcef6c19-serving-cert\") pod \"48e63397-f90d-4325-b5ce-4e96dcef6c19\" (UID: \"48e63397-f90d-4325-b5ce-4e96dcef6c19\") " Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.000544 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/48e63397-f90d-4325-b5ce-4e96dcef6c19-client-ca\") pod \"48e63397-f90d-4325-b5ce-4e96dcef6c19\" (UID: \"48e63397-f90d-4325-b5ce-4e96dcef6c19\") " Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.000581 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btlqj\" (UniqueName: \"kubernetes.io/projected/48e63397-f90d-4325-b5ce-4e96dcef6c19-kube-api-access-btlqj\") pod \"48e63397-f90d-4325-b5ce-4e96dcef6c19\" (UID: \"48e63397-f90d-4325-b5ce-4e96dcef6c19\") " Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.001698 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48e63397-f90d-4325-b5ce-4e96dcef6c19-config" (OuterVolumeSpecName: "config") pod "48e63397-f90d-4325-b5ce-4e96dcef6c19" (UID: "48e63397-f90d-4325-b5ce-4e96dcef6c19"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.001716 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48e63397-f90d-4325-b5ce-4e96dcef6c19-client-ca" (OuterVolumeSpecName: "client-ca") pod "48e63397-f90d-4325-b5ce-4e96dcef6c19" (UID: "48e63397-f90d-4325-b5ce-4e96dcef6c19"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.006533 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e63397-f90d-4325-b5ce-4e96dcef6c19-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "48e63397-f90d-4325-b5ce-4e96dcef6c19" (UID: "48e63397-f90d-4325-b5ce-4e96dcef6c19"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.007066 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48e63397-f90d-4325-b5ce-4e96dcef6c19-kube-api-access-btlqj" (OuterVolumeSpecName: "kube-api-access-btlqj") pod "48e63397-f90d-4325-b5ce-4e96dcef6c19" (UID: "48e63397-f90d-4325-b5ce-4e96dcef6c19"). InnerVolumeSpecName "kube-api-access-btlqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.103185 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48e63397-f90d-4325-b5ce-4e96dcef6c19-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.103239 4806 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48e63397-f90d-4325-b5ce-4e96dcef6c19-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.103260 4806 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/48e63397-f90d-4325-b5ce-4e96dcef6c19-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.103273 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btlqj\" (UniqueName: \"kubernetes.io/projected/48e63397-f90d-4325-b5ce-4e96dcef6c19-kube-api-access-btlqj\") on node \"crc\" DevicePath \"\"" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.434784 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" event={"ID":"48e63397-f90d-4325-b5ce-4e96dcef6c19","Type":"ContainerDied","Data":"24c6fc1a15a5d7174b521f35cc341a2e0e5f3f72e2266a541e7c5dc13f1187c7"} Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.434893 4806 scope.go:117] "RemoveContainer" containerID="20ece38be669e0adb340a882012a0299d4fc0a0106e4106917fa07580ba6783b" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.434841 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.477504 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4"] Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.480991 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-795597c8b4-m75l4"] Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.558818 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww"] Dec 04 04:01:16 crc kubenswrapper[4806]: E1204 04:01:16.559276 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48e63397-f90d-4325-b5ce-4e96dcef6c19" containerName="route-controller-manager" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.559297 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="48e63397-f90d-4325-b5ce-4e96dcef6c19" containerName="route-controller-manager" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.559436 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="48e63397-f90d-4325-b5ce-4e96dcef6c19" containerName="route-controller-manager" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.560200 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.562822 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.562822 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.563990 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.564105 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.564001 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.564010 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.570323 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww"] Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.611423 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53cbc089-5334-46dd-9d5d-361183f9e999-config\") pod \"route-controller-manager-749699db6b-c92ww\" (UID: \"53cbc089-5334-46dd-9d5d-361183f9e999\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.611464 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rc54\" (UniqueName: \"kubernetes.io/projected/53cbc089-5334-46dd-9d5d-361183f9e999-kube-api-access-5rc54\") pod \"route-controller-manager-749699db6b-c92ww\" (UID: \"53cbc089-5334-46dd-9d5d-361183f9e999\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.611509 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53cbc089-5334-46dd-9d5d-361183f9e999-client-ca\") pod \"route-controller-manager-749699db6b-c92ww\" (UID: \"53cbc089-5334-46dd-9d5d-361183f9e999\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.611614 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53cbc089-5334-46dd-9d5d-361183f9e999-serving-cert\") pod \"route-controller-manager-749699db6b-c92ww\" (UID: \"53cbc089-5334-46dd-9d5d-361183f9e999\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.714069 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53cbc089-5334-46dd-9d5d-361183f9e999-config\") pod \"route-controller-manager-749699db6b-c92ww\" (UID: \"53cbc089-5334-46dd-9d5d-361183f9e999\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.714690 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rc54\" (UniqueName: \"kubernetes.io/projected/53cbc089-5334-46dd-9d5d-361183f9e999-kube-api-access-5rc54\") pod \"route-controller-manager-749699db6b-c92ww\" (UID: \"53cbc089-5334-46dd-9d5d-361183f9e999\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.714887 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53cbc089-5334-46dd-9d5d-361183f9e999-client-ca\") pod \"route-controller-manager-749699db6b-c92ww\" (UID: \"53cbc089-5334-46dd-9d5d-361183f9e999\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.715076 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53cbc089-5334-46dd-9d5d-361183f9e999-serving-cert\") pod \"route-controller-manager-749699db6b-c92ww\" (UID: \"53cbc089-5334-46dd-9d5d-361183f9e999\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.715720 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53cbc089-5334-46dd-9d5d-361183f9e999-config\") pod \"route-controller-manager-749699db6b-c92ww\" (UID: \"53cbc089-5334-46dd-9d5d-361183f9e999\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.716361 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53cbc089-5334-46dd-9d5d-361183f9e999-client-ca\") pod \"route-controller-manager-749699db6b-c92ww\" (UID: \"53cbc089-5334-46dd-9d5d-361183f9e999\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.719043 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53cbc089-5334-46dd-9d5d-361183f9e999-serving-cert\") pod \"route-controller-manager-749699db6b-c92ww\" (UID: \"53cbc089-5334-46dd-9d5d-361183f9e999\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.736283 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rc54\" (UniqueName: \"kubernetes.io/projected/53cbc089-5334-46dd-9d5d-361183f9e999-kube-api-access-5rc54\") pod \"route-controller-manager-749699db6b-c92ww\" (UID: \"53cbc089-5334-46dd-9d5d-361183f9e999\") " pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" Dec 04 04:01:16 crc kubenswrapper[4806]: I1204 04:01:16.895696 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" Dec 04 04:01:17 crc kubenswrapper[4806]: I1204 04:01:17.332343 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww"] Dec 04 04:01:17 crc kubenswrapper[4806]: I1204 04:01:17.432484 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48e63397-f90d-4325-b5ce-4e96dcef6c19" path="/var/lib/kubelet/pods/48e63397-f90d-4325-b5ce-4e96dcef6c19/volumes" Dec 04 04:01:17 crc kubenswrapper[4806]: I1204 04:01:17.444493 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" event={"ID":"53cbc089-5334-46dd-9d5d-361183f9e999","Type":"ContainerStarted","Data":"4bd7558c3234ecc0b17adba54b6b605cd94c7ee0330c677d27c3dc8aee2753eb"} Dec 04 04:01:18 crc kubenswrapper[4806]: I1204 04:01:18.451219 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" event={"ID":"53cbc089-5334-46dd-9d5d-361183f9e999","Type":"ContainerStarted","Data":"48338ab15a030226e5109b2c4d896390f7df0f14dc44314b6122e9f12525741c"} Dec 04 04:01:18 crc kubenswrapper[4806]: I1204 04:01:18.451622 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" Dec 04 04:01:18 crc kubenswrapper[4806]: I1204 04:01:18.458137 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" Dec 04 04:01:18 crc kubenswrapper[4806]: I1204 04:01:18.478273 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-749699db6b-c92ww" podStartSLOduration=4.478235581 podStartE2EDuration="4.478235581s" podCreationTimestamp="2025-12-04 04:01:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:01:18.471148281 +0000 UTC m=+393.329661209" watchObservedRunningTime="2025-12-04 04:01:18.478235581 +0000 UTC m=+393.336748509" Dec 04 04:01:22 crc kubenswrapper[4806]: I1204 04:01:22.885870 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-9vsks" Dec 04 04:01:22 crc kubenswrapper[4806]: I1204 04:01:22.962816 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c2d5j"] Dec 04 04:01:27 crc kubenswrapper[4806]: I1204 04:01:27.047398 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:01:27 crc kubenswrapper[4806]: I1204 04:01:27.047881 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:01:27 crc kubenswrapper[4806]: I1204 04:01:27.047959 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 04:01:27 crc kubenswrapper[4806]: I1204 04:01:27.048744 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3e4ceeaa6204e4c1f2d29168fb61b4f1c4c0b7afb94d21facd0f0961ea7e32ff"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 04:01:27 crc kubenswrapper[4806]: I1204 04:01:27.048808 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://3e4ceeaa6204e4c1f2d29168fb61b4f1c4c0b7afb94d21facd0f0961ea7e32ff" gracePeriod=600 Dec 04 04:01:27 crc kubenswrapper[4806]: I1204 04:01:27.513095 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="3e4ceeaa6204e4c1f2d29168fb61b4f1c4c0b7afb94d21facd0f0961ea7e32ff" exitCode=0 Dec 04 04:01:27 crc kubenswrapper[4806]: I1204 04:01:27.513545 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"3e4ceeaa6204e4c1f2d29168fb61b4f1c4c0b7afb94d21facd0f0961ea7e32ff"} Dec 04 04:01:27 crc kubenswrapper[4806]: I1204 04:01:27.513585 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"14fd61c0a81ae92a7b878dd7a4719064b58b4a7d3bf90d5f039134a7e74c54ac"} Dec 04 04:01:27 crc kubenswrapper[4806]: I1204 04:01:27.513609 4806 scope.go:117] "RemoveContainer" containerID="c128970289043f42190fb1bc83eba96b715b2e86b1ed3a7a2dffdc0f3d90844d" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.010576 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" podUID="233ad320-ca2f-40bf-bdcf-64e5aed5509f" containerName="registry" containerID="cri-o://2501b5edc73f99907bc36c581f839bfc572d15b6f6a8af53682d41fdf18d85c9" gracePeriod=30 Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.427644 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.580295 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-bound-sa-token\") pod \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.580542 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.580585 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/233ad320-ca2f-40bf-bdcf-64e5aed5509f-registry-certificates\") pod \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.580636 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/233ad320-ca2f-40bf-bdcf-64e5aed5509f-installation-pull-secrets\") pod \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.580681 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dfts\" (UniqueName: \"kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-kube-api-access-4dfts\") pod \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.580713 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/233ad320-ca2f-40bf-bdcf-64e5aed5509f-ca-trust-extracted\") pod \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.580764 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-registry-tls\") pod \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.580799 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/233ad320-ca2f-40bf-bdcf-64e5aed5509f-trusted-ca\") pod \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\" (UID: \"233ad320-ca2f-40bf-bdcf-64e5aed5509f\") " Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.581714 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/233ad320-ca2f-40bf-bdcf-64e5aed5509f-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "233ad320-ca2f-40bf-bdcf-64e5aed5509f" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.581755 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/233ad320-ca2f-40bf-bdcf-64e5aed5509f-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "233ad320-ca2f-40bf-bdcf-64e5aed5509f" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.588018 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/233ad320-ca2f-40bf-bdcf-64e5aed5509f-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "233ad320-ca2f-40bf-bdcf-64e5aed5509f" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.588618 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "233ad320-ca2f-40bf-bdcf-64e5aed5509f" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.598018 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "233ad320-ca2f-40bf-bdcf-64e5aed5509f" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.599475 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-kube-api-access-4dfts" (OuterVolumeSpecName: "kube-api-access-4dfts") pod "233ad320-ca2f-40bf-bdcf-64e5aed5509f" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f"). InnerVolumeSpecName "kube-api-access-4dfts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.599525 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "233ad320-ca2f-40bf-bdcf-64e5aed5509f" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.611154 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/233ad320-ca2f-40bf-bdcf-64e5aed5509f-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "233ad320-ca2f-40bf-bdcf-64e5aed5509f" (UID: "233ad320-ca2f-40bf-bdcf-64e5aed5509f"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.681839 4806 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/233ad320-ca2f-40bf-bdcf-64e5aed5509f-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.681876 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dfts\" (UniqueName: \"kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-kube-api-access-4dfts\") on node \"crc\" DevicePath \"\"" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.681887 4806 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/233ad320-ca2f-40bf-bdcf-64e5aed5509f-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.681898 4806 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.681908 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/233ad320-ca2f-40bf-bdcf-64e5aed5509f-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.681917 4806 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/233ad320-ca2f-40bf-bdcf-64e5aed5509f-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.681943 4806 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/233ad320-ca2f-40bf-bdcf-64e5aed5509f-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.927543 4806 generic.go:334] "Generic (PLEG): container finished" podID="233ad320-ca2f-40bf-bdcf-64e5aed5509f" containerID="2501b5edc73f99907bc36c581f839bfc572d15b6f6a8af53682d41fdf18d85c9" exitCode=0 Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.927628 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.927614 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" event={"ID":"233ad320-ca2f-40bf-bdcf-64e5aed5509f","Type":"ContainerDied","Data":"2501b5edc73f99907bc36c581f839bfc572d15b6f6a8af53682d41fdf18d85c9"} Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.927703 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c2d5j" event={"ID":"233ad320-ca2f-40bf-bdcf-64e5aed5509f","Type":"ContainerDied","Data":"e6d5c3fe8047b8d92915969a5f24a5965881126ac40f73d758137e4f2d95b3fd"} Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.927733 4806 scope.go:117] "RemoveContainer" containerID="2501b5edc73f99907bc36c581f839bfc572d15b6f6a8af53682d41fdf18d85c9" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.957429 4806 scope.go:117] "RemoveContainer" containerID="2501b5edc73f99907bc36c581f839bfc572d15b6f6a8af53682d41fdf18d85c9" Dec 04 04:01:48 crc kubenswrapper[4806]: E1204 04:01:48.958254 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2501b5edc73f99907bc36c581f839bfc572d15b6f6a8af53682d41fdf18d85c9\": container with ID starting with 2501b5edc73f99907bc36c581f839bfc572d15b6f6a8af53682d41fdf18d85c9 not found: ID does not exist" containerID="2501b5edc73f99907bc36c581f839bfc572d15b6f6a8af53682d41fdf18d85c9" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.958299 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2501b5edc73f99907bc36c581f839bfc572d15b6f6a8af53682d41fdf18d85c9"} err="failed to get container status \"2501b5edc73f99907bc36c581f839bfc572d15b6f6a8af53682d41fdf18d85c9\": rpc error: code = NotFound desc = could not find container \"2501b5edc73f99907bc36c581f839bfc572d15b6f6a8af53682d41fdf18d85c9\": container with ID starting with 2501b5edc73f99907bc36c581f839bfc572d15b6f6a8af53682d41fdf18d85c9 not found: ID does not exist" Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.966374 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c2d5j"] Dec 04 04:01:48 crc kubenswrapper[4806]: I1204 04:01:48.973694 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c2d5j"] Dec 04 04:01:49 crc kubenswrapper[4806]: I1204 04:01:49.431226 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="233ad320-ca2f-40bf-bdcf-64e5aed5509f" path="/var/lib/kubelet/pods/233ad320-ca2f-40bf-bdcf-64e5aed5509f/volumes" Dec 04 04:03:27 crc kubenswrapper[4806]: I1204 04:03:27.048236 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:03:27 crc kubenswrapper[4806]: I1204 04:03:27.049223 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:03:57 crc kubenswrapper[4806]: I1204 04:03:57.048067 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:03:57 crc kubenswrapper[4806]: I1204 04:03:57.049028 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:04:27 crc kubenswrapper[4806]: I1204 04:04:27.047865 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:04:27 crc kubenswrapper[4806]: I1204 04:04:27.048991 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:04:27 crc kubenswrapper[4806]: I1204 04:04:27.049069 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 04:04:27 crc kubenswrapper[4806]: I1204 04:04:27.049985 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"14fd61c0a81ae92a7b878dd7a4719064b58b4a7d3bf90d5f039134a7e74c54ac"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 04:04:27 crc kubenswrapper[4806]: I1204 04:04:27.050050 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://14fd61c0a81ae92a7b878dd7a4719064b58b4a7d3bf90d5f039134a7e74c54ac" gracePeriod=600 Dec 04 04:04:27 crc kubenswrapper[4806]: I1204 04:04:27.952966 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="14fd61c0a81ae92a7b878dd7a4719064b58b4a7d3bf90d5f039134a7e74c54ac" exitCode=0 Dec 04 04:04:27 crc kubenswrapper[4806]: I1204 04:04:27.953066 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"14fd61c0a81ae92a7b878dd7a4719064b58b4a7d3bf90d5f039134a7e74c54ac"} Dec 04 04:04:27 crc kubenswrapper[4806]: I1204 04:04:27.953784 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"f0bc0378fad2cb0fb66c2e5ff86d10f9a724302404985dba5e351fcfd4397778"} Dec 04 04:04:27 crc kubenswrapper[4806]: I1204 04:04:27.953819 4806 scope.go:117] "RemoveContainer" containerID="3e4ceeaa6204e4c1f2d29168fb61b4f1c4c0b7afb94d21facd0f0961ea7e32ff" Dec 04 04:06:18 crc kubenswrapper[4806]: I1204 04:06:18.907903 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-8vl9s"] Dec 04 04:06:18 crc kubenswrapper[4806]: E1204 04:06:18.909124 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="233ad320-ca2f-40bf-bdcf-64e5aed5509f" containerName="registry" Dec 04 04:06:18 crc kubenswrapper[4806]: I1204 04:06:18.909143 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="233ad320-ca2f-40bf-bdcf-64e5aed5509f" containerName="registry" Dec 04 04:06:18 crc kubenswrapper[4806]: I1204 04:06:18.909297 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="233ad320-ca2f-40bf-bdcf-64e5aed5509f" containerName="registry" Dec 04 04:06:18 crc kubenswrapper[4806]: I1204 04:06:18.909907 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-8vl9s" Dec 04 04:06:18 crc kubenswrapper[4806]: I1204 04:06:18.915734 4806 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-bflwt" Dec 04 04:06:18 crc kubenswrapper[4806]: I1204 04:06:18.916053 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 04 04:06:18 crc kubenswrapper[4806]: I1204 04:06:18.921570 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 04 04:06:18 crc kubenswrapper[4806]: I1204 04:06:18.939263 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-8vl9s"] Dec 04 04:06:18 crc kubenswrapper[4806]: I1204 04:06:18.945530 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-6l44d"] Dec 04 04:06:18 crc kubenswrapper[4806]: I1204 04:06:18.946607 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-6l44d" Dec 04 04:06:18 crc kubenswrapper[4806]: I1204 04:06:18.954497 4806 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-l4pnb" Dec 04 04:06:18 crc kubenswrapper[4806]: I1204 04:06:18.961298 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-dzbcz"] Dec 04 04:06:18 crc kubenswrapper[4806]: I1204 04:06:18.962266 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-dzbcz" Dec 04 04:06:18 crc kubenswrapper[4806]: I1204 04:06:18.969266 4806 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-kb2w9" Dec 04 04:06:18 crc kubenswrapper[4806]: I1204 04:06:18.975076 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-6l44d"] Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.005537 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-dzbcz"] Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.050351 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gktdk\" (UniqueName: \"kubernetes.io/projected/ed32bab6-fee8-43d7-a157-9e8d5d17b98a-kube-api-access-gktdk\") pod \"cert-manager-5b446d88c5-6l44d\" (UID: \"ed32bab6-fee8-43d7-a157-9e8d5d17b98a\") " pod="cert-manager/cert-manager-5b446d88c5-6l44d" Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.050636 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjq4s\" (UniqueName: \"kubernetes.io/projected/65faa3e8-8810-4a0e-a3ec-bb2b4d6283fb-kube-api-access-gjq4s\") pod \"cert-manager-cainjector-7f985d654d-8vl9s\" (UID: \"65faa3e8-8810-4a0e-a3ec-bb2b4d6283fb\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-8vl9s" Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.152030 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjq4s\" (UniqueName: \"kubernetes.io/projected/65faa3e8-8810-4a0e-a3ec-bb2b4d6283fb-kube-api-access-gjq4s\") pod \"cert-manager-cainjector-7f985d654d-8vl9s\" (UID: \"65faa3e8-8810-4a0e-a3ec-bb2b4d6283fb\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-8vl9s" Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.152103 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk8zr\" (UniqueName: \"kubernetes.io/projected/7934a456-a3ec-4a88-8857-17d7f7db5837-kube-api-access-pk8zr\") pod \"cert-manager-webhook-5655c58dd6-dzbcz\" (UID: \"7934a456-a3ec-4a88-8857-17d7f7db5837\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-dzbcz" Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.152145 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gktdk\" (UniqueName: \"kubernetes.io/projected/ed32bab6-fee8-43d7-a157-9e8d5d17b98a-kube-api-access-gktdk\") pod \"cert-manager-5b446d88c5-6l44d\" (UID: \"ed32bab6-fee8-43d7-a157-9e8d5d17b98a\") " pod="cert-manager/cert-manager-5b446d88c5-6l44d" Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.175731 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjq4s\" (UniqueName: \"kubernetes.io/projected/65faa3e8-8810-4a0e-a3ec-bb2b4d6283fb-kube-api-access-gjq4s\") pod \"cert-manager-cainjector-7f985d654d-8vl9s\" (UID: \"65faa3e8-8810-4a0e-a3ec-bb2b4d6283fb\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-8vl9s" Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.175731 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gktdk\" (UniqueName: \"kubernetes.io/projected/ed32bab6-fee8-43d7-a157-9e8d5d17b98a-kube-api-access-gktdk\") pod \"cert-manager-5b446d88c5-6l44d\" (UID: \"ed32bab6-fee8-43d7-a157-9e8d5d17b98a\") " pod="cert-manager/cert-manager-5b446d88c5-6l44d" Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.243745 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-8vl9s" Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.253670 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk8zr\" (UniqueName: \"kubernetes.io/projected/7934a456-a3ec-4a88-8857-17d7f7db5837-kube-api-access-pk8zr\") pod \"cert-manager-webhook-5655c58dd6-dzbcz\" (UID: \"7934a456-a3ec-4a88-8857-17d7f7db5837\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-dzbcz" Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.262478 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-6l44d" Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.279236 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk8zr\" (UniqueName: \"kubernetes.io/projected/7934a456-a3ec-4a88-8857-17d7f7db5837-kube-api-access-pk8zr\") pod \"cert-manager-webhook-5655c58dd6-dzbcz\" (UID: \"7934a456-a3ec-4a88-8857-17d7f7db5837\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-dzbcz" Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.279739 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-dzbcz" Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.530593 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-8vl9s"] Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.545700 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.614395 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-dzbcz"] Dec 04 04:06:19 crc kubenswrapper[4806]: W1204 04:06:19.622305 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7934a456_a3ec_4a88_8857_17d7f7db5837.slice/crio-e47c4974b2344869846d264095d587a9fa0cd248554cac0101a66bee2b2cd6f6 WatchSource:0}: Error finding container e47c4974b2344869846d264095d587a9fa0cd248554cac0101a66bee2b2cd6f6: Status 404 returned error can't find the container with id e47c4974b2344869846d264095d587a9fa0cd248554cac0101a66bee2b2cd6f6 Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.661738 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-8vl9s" event={"ID":"65faa3e8-8810-4a0e-a3ec-bb2b4d6283fb","Type":"ContainerStarted","Data":"f9053c5167e494463cb5f8f4c7f8a9d493c77922cdcde4f280696e2525d22c66"} Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.663818 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-dzbcz" event={"ID":"7934a456-a3ec-4a88-8857-17d7f7db5837","Type":"ContainerStarted","Data":"e47c4974b2344869846d264095d587a9fa0cd248554cac0101a66bee2b2cd6f6"} Dec 04 04:06:19 crc kubenswrapper[4806]: I1204 04:06:19.672774 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-6l44d"] Dec 04 04:06:20 crc kubenswrapper[4806]: I1204 04:06:20.677850 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-6l44d" event={"ID":"ed32bab6-fee8-43d7-a157-9e8d5d17b98a","Type":"ContainerStarted","Data":"ba836b66a2512dbc1854efc89c21024ac690b28257392066c1def2d17d76eb1d"} Dec 04 04:06:23 crc kubenswrapper[4806]: I1204 04:06:23.702020 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-dzbcz" event={"ID":"7934a456-a3ec-4a88-8857-17d7f7db5837","Type":"ContainerStarted","Data":"7301b052258f0efde893a0360c4689bbbc034bfad500406c1dadc75eae08b82c"} Dec 04 04:06:23 crc kubenswrapper[4806]: I1204 04:06:23.703134 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-dzbcz" Dec 04 04:06:23 crc kubenswrapper[4806]: I1204 04:06:23.704462 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-6l44d" event={"ID":"ed32bab6-fee8-43d7-a157-9e8d5d17b98a","Type":"ContainerStarted","Data":"d09ccda239d84e6298b4f428c43b96bb34f1861bb399a5e4a0e2615c4cccd892"} Dec 04 04:06:23 crc kubenswrapper[4806]: I1204 04:06:23.706310 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-8vl9s" event={"ID":"65faa3e8-8810-4a0e-a3ec-bb2b4d6283fb","Type":"ContainerStarted","Data":"4165d6b7cf06fa3da2c6927908f821251655b2254ff4453f3784181759a8e194"} Dec 04 04:06:23 crc kubenswrapper[4806]: I1204 04:06:23.723270 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-dzbcz" podStartSLOduration=2.621110867 podStartE2EDuration="5.723245141s" podCreationTimestamp="2025-12-04 04:06:18 +0000 UTC" firstStartedPulling="2025-12-04 04:06:19.627864151 +0000 UTC m=+694.486377079" lastFinishedPulling="2025-12-04 04:06:22.729998425 +0000 UTC m=+697.588511353" observedRunningTime="2025-12-04 04:06:23.720241187 +0000 UTC m=+698.578754115" watchObservedRunningTime="2025-12-04 04:06:23.723245141 +0000 UTC m=+698.581758069" Dec 04 04:06:23 crc kubenswrapper[4806]: I1204 04:06:23.741793 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-6l44d" podStartSLOduration=2.541811624 podStartE2EDuration="5.741760921s" podCreationTimestamp="2025-12-04 04:06:18 +0000 UTC" firstStartedPulling="2025-12-04 04:06:19.680960994 +0000 UTC m=+694.539473922" lastFinishedPulling="2025-12-04 04:06:22.880910291 +0000 UTC m=+697.739423219" observedRunningTime="2025-12-04 04:06:23.74043164 +0000 UTC m=+698.598944568" watchObservedRunningTime="2025-12-04 04:06:23.741760921 +0000 UTC m=+698.600273859" Dec 04 04:06:27 crc kubenswrapper[4806]: I1204 04:06:27.047360 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:06:27 crc kubenswrapper[4806]: I1204 04:06:27.048378 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:06:29 crc kubenswrapper[4806]: I1204 04:06:29.283338 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-dzbcz" Dec 04 04:06:29 crc kubenswrapper[4806]: I1204 04:06:29.303084 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-8vl9s" podStartSLOduration=8.037421177 podStartE2EDuration="11.303058752s" podCreationTimestamp="2025-12-04 04:06:18 +0000 UTC" firstStartedPulling="2025-12-04 04:06:19.545378427 +0000 UTC m=+694.403891355" lastFinishedPulling="2025-12-04 04:06:22.811016002 +0000 UTC m=+697.669528930" observedRunningTime="2025-12-04 04:06:23.76567139 +0000 UTC m=+698.624184318" watchObservedRunningTime="2025-12-04 04:06:29.303058752 +0000 UTC m=+704.161571680" Dec 04 04:06:38 crc kubenswrapper[4806]: I1204 04:06:38.662587 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ljxsg"] Dec 04 04:06:38 crc kubenswrapper[4806]: I1204 04:06:38.664130 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovn-controller" containerID="cri-o://dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28" gracePeriod=30 Dec 04 04:06:38 crc kubenswrapper[4806]: I1204 04:06:38.664857 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="sbdb" containerID="cri-o://e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546" gracePeriod=30 Dec 04 04:06:38 crc kubenswrapper[4806]: I1204 04:06:38.664919 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="nbdb" containerID="cri-o://24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7" gracePeriod=30 Dec 04 04:06:38 crc kubenswrapper[4806]: I1204 04:06:38.664997 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="northd" containerID="cri-o://ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7" gracePeriod=30 Dec 04 04:06:38 crc kubenswrapper[4806]: I1204 04:06:38.665068 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c" gracePeriod=30 Dec 04 04:06:38 crc kubenswrapper[4806]: I1204 04:06:38.665131 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="kube-rbac-proxy-node" containerID="cri-o://026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b" gracePeriod=30 Dec 04 04:06:38 crc kubenswrapper[4806]: I1204 04:06:38.665193 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovn-acl-logging" containerID="cri-o://5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526" gracePeriod=30 Dec 04 04:06:38 crc kubenswrapper[4806]: I1204 04:06:38.703741 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovnkube-controller" containerID="cri-o://9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8" gracePeriod=30 Dec 04 04:06:38 crc kubenswrapper[4806]: I1204 04:06:38.815071 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dsclk_4f13bcfc-c287-40fd-b8f6-e4bbea8c6577/kube-multus/2.log" Dec 04 04:06:38 crc kubenswrapper[4806]: I1204 04:06:38.816916 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dsclk_4f13bcfc-c287-40fd-b8f6-e4bbea8c6577/kube-multus/1.log" Dec 04 04:06:38 crc kubenswrapper[4806]: I1204 04:06:38.817747 4806 generic.go:334] "Generic (PLEG): container finished" podID="4f13bcfc-c287-40fd-b8f6-e4bbea8c6577" containerID="d6efe030cffcc9ae048a86abeffad5784f713909803cbabe74db70e9e5212015" exitCode=2 Dec 04 04:06:38 crc kubenswrapper[4806]: I1204 04:06:38.817798 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dsclk" event={"ID":"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577","Type":"ContainerDied","Data":"d6efe030cffcc9ae048a86abeffad5784f713909803cbabe74db70e9e5212015"} Dec 04 04:06:38 crc kubenswrapper[4806]: I1204 04:06:38.817851 4806 scope.go:117] "RemoveContainer" containerID="a6e29eb2e4c712cf0fcf618844fe548f0c38982e08169a900c816f7a9a6b71ba" Dec 04 04:06:38 crc kubenswrapper[4806]: I1204 04:06:38.818672 4806 scope.go:117] "RemoveContainer" containerID="d6efe030cffcc9ae048a86abeffad5784f713909803cbabe74db70e9e5212015" Dec 04 04:06:38 crc kubenswrapper[4806]: E1204 04:06:38.819156 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-dsclk_openshift-multus(4f13bcfc-c287-40fd-b8f6-e4bbea8c6577)\"" pod="openshift-multus/multus-dsclk" podUID="4f13bcfc-c287-40fd-b8f6-e4bbea8c6577" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.047340 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/4.log" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.048362 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/3.log" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.051481 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovn-acl-logging/0.log" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.052309 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovn-controller/0.log" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.053048 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.121204 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5cqmh"] Dec 04 04:06:39 crc kubenswrapper[4806]: E1204 04:06:39.122074 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="sbdb" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122100 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="sbdb" Dec 04 04:06:39 crc kubenswrapper[4806]: E1204 04:06:39.122142 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovnkube-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122153 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovnkube-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: E1204 04:06:39.122165 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovn-acl-logging" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122175 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovn-acl-logging" Dec 04 04:06:39 crc kubenswrapper[4806]: E1204 04:06:39.122190 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122200 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 04:06:39 crc kubenswrapper[4806]: E1204 04:06:39.122236 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovn-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122246 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovn-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: E1204 04:06:39.122271 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="nbdb" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122280 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="nbdb" Dec 04 04:06:39 crc kubenswrapper[4806]: E1204 04:06:39.122299 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="northd" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122308 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="northd" Dec 04 04:06:39 crc kubenswrapper[4806]: E1204 04:06:39.122332 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="kubecfg-setup" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122343 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="kubecfg-setup" Dec 04 04:06:39 crc kubenswrapper[4806]: E1204 04:06:39.122352 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="kube-rbac-proxy-node" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122362 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="kube-rbac-proxy-node" Dec 04 04:06:39 crc kubenswrapper[4806]: E1204 04:06:39.122372 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovnkube-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122382 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovnkube-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: E1204 04:06:39.122399 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovnkube-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122408 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovnkube-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: E1204 04:06:39.122424 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovnkube-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122433 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovnkube-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122751 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovn-acl-logging" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122772 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122803 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="nbdb" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122822 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovnkube-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122835 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovnkube-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122849 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovnkube-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122867 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovnkube-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122881 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="kube-rbac-proxy-node" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122898 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="northd" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122913 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="sbdb" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.122940 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovn-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: E1204 04:06:39.123274 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovnkube-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.123295 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovnkube-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.123613 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerName="ovnkube-controller" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.129153 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.172555 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-slash\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.172612 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-cni-bin\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.172651 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovnkube-config\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.172651 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-slash" (OuterVolumeSpecName: "host-slash") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.172678 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-openvswitch\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.172698 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-run-netns\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.172724 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-var-lib-openvswitch\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.172758 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovn-node-metrics-cert\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.172783 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-ovn\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.172811 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-node-log\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.172834 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-run-ovn-kubernetes\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.172861 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-log-socket\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.172894 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-systemd\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.172984 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-var-lib-cni-networks-ovn-kubernetes\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173009 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovnkube-script-lib\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173036 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8ccw\" (UniqueName: \"kubernetes.io/projected/087077d3-f9fa-4d16-aa7c-eb4241cbb293-kube-api-access-j8ccw\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173057 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-systemd-units\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173087 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-env-overrides\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173114 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-kubelet\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173138 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-etc-openvswitch\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173177 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-cni-netd\") pod \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\" (UID: \"087077d3-f9fa-4d16-aa7c-eb4241cbb293\") " Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.172699 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173212 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173241 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173259 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173275 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173305 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173297 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-log-socket" (OuterVolumeSpecName: "log-socket") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173349 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-etc-openvswitch\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173321 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173375 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-run-netns\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173398 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-run-openvswitch\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173420 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-run-ovn\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173451 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-var-lib-openvswitch\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173487 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-node-log\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173513 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b18e6557-daa1-469e-9042-7fcc2d2d5140-env-overrides\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173535 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-cni-netd\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173562 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-kubelet\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173589 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-run-ovn-kubernetes\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173620 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-slash\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173644 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-systemd-units\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173671 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b18e6557-daa1-469e-9042-7fcc2d2d5140-ovn-node-metrics-cert\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173699 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-log-socket\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173726 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b18e6557-daa1-469e-9042-7fcc2d2d5140-ovnkube-script-lib\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173749 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9xj2\" (UniqueName: \"kubernetes.io/projected/b18e6557-daa1-469e-9042-7fcc2d2d5140-kube-api-access-w9xj2\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173775 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-run-systemd\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173802 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-cni-bin\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173824 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b18e6557-daa1-469e-9042-7fcc2d2d5140-ovnkube-config\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173873 4806 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-slash\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173890 4806 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173904 4806 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173916 4806 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173946 4806 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173959 4806 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173971 4806 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173983 4806 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-log-socket\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173377 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-node-log" (OuterVolumeSpecName: "node-log") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173448 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.173475 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.174047 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.175325 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.175476 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.175502 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.176160 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.177079 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.181164 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.181208 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/087077d3-f9fa-4d16-aa7c-eb4241cbb293-kube-api-access-j8ccw" (OuterVolumeSpecName: "kube-api-access-j8ccw") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "kube-api-access-j8ccw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.189269 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "087077d3-f9fa-4d16-aa7c-eb4241cbb293" (UID: "087077d3-f9fa-4d16-aa7c-eb4241cbb293"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274296 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-node-log\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274362 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b18e6557-daa1-469e-9042-7fcc2d2d5140-env-overrides\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274384 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-cni-netd\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274407 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-kubelet\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274430 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-run-ovn-kubernetes\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274452 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-slash\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274470 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-systemd-units\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274488 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b18e6557-daa1-469e-9042-7fcc2d2d5140-ovn-node-metrics-cert\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274505 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-log-socket\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274520 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b18e6557-daa1-469e-9042-7fcc2d2d5140-ovnkube-script-lib\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274538 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-run-systemd\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274557 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9xj2\" (UniqueName: \"kubernetes.io/projected/b18e6557-daa1-469e-9042-7fcc2d2d5140-kube-api-access-w9xj2\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274591 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-cni-bin\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274624 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b18e6557-daa1-469e-9042-7fcc2d2d5140-ovnkube-config\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274655 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274694 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-etc-openvswitch\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274723 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-run-netns\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274745 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-run-openvswitch\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274766 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-run-ovn\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.274789 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-var-lib-openvswitch\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275034 4806 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275057 4806 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275073 4806 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-node-log\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275090 4806 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275127 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-node-log\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275382 4806 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275507 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-run-systemd\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275549 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-systemd-units\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275591 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-slash\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275653 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-cni-netd\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275681 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-kubelet\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275704 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-run-ovn-kubernetes\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275741 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-etc-openvswitch\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275762 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-cni-bin\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275917 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-run-openvswitch\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275972 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-run-ovn\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.275998 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-run-netns\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.276035 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-log-socket\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.276075 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-var-lib-openvswitch\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.276112 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b18e6557-daa1-469e-9042-7fcc2d2d5140-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.276142 4806 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.276168 4806 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.276184 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8ccw\" (UniqueName: \"kubernetes.io/projected/087077d3-f9fa-4d16-aa7c-eb4241cbb293-kube-api-access-j8ccw\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.276198 4806 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.276211 4806 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/087077d3-f9fa-4d16-aa7c-eb4241cbb293-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.276265 4806 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.276283 4806 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/087077d3-f9fa-4d16-aa7c-eb4241cbb293-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.276768 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b18e6557-daa1-469e-9042-7fcc2d2d5140-env-overrides\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.277079 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b18e6557-daa1-469e-9042-7fcc2d2d5140-ovnkube-config\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.277201 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b18e6557-daa1-469e-9042-7fcc2d2d5140-ovnkube-script-lib\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.279760 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b18e6557-daa1-469e-9042-7fcc2d2d5140-ovn-node-metrics-cert\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.295285 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9xj2\" (UniqueName: \"kubernetes.io/projected/b18e6557-daa1-469e-9042-7fcc2d2d5140-kube-api-access-w9xj2\") pod \"ovnkube-node-5cqmh\" (UID: \"b18e6557-daa1-469e-9042-7fcc2d2d5140\") " pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.448517 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:39 crc kubenswrapper[4806]: W1204 04:06:39.475080 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb18e6557_daa1_469e_9042_7fcc2d2d5140.slice/crio-5fd5663e57b71930d0f411d197deb4154fbf482218a1cc390fbbd609c66aa353 WatchSource:0}: Error finding container 5fd5663e57b71930d0f411d197deb4154fbf482218a1cc390fbbd609c66aa353: Status 404 returned error can't find the container with id 5fd5663e57b71930d0f411d197deb4154fbf482218a1cc390fbbd609c66aa353 Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.827870 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dsclk_4f13bcfc-c287-40fd-b8f6-e4bbea8c6577/kube-multus/2.log" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.829826 4806 generic.go:334] "Generic (PLEG): container finished" podID="b18e6557-daa1-469e-9042-7fcc2d2d5140" containerID="0720dbc99a204695ae3e2eee63cc15680f7daacfbb9ced8ef68e18819e101036" exitCode=0 Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.829897 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" event={"ID":"b18e6557-daa1-469e-9042-7fcc2d2d5140","Type":"ContainerDied","Data":"0720dbc99a204695ae3e2eee63cc15680f7daacfbb9ced8ef68e18819e101036"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.829971 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" event={"ID":"b18e6557-daa1-469e-9042-7fcc2d2d5140","Type":"ContainerStarted","Data":"5fd5663e57b71930d0f411d197deb4154fbf482218a1cc390fbbd609c66aa353"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.835606 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/4.log" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.839104 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovnkube-controller/3.log" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.844902 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovn-acl-logging/0.log" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.845403 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ljxsg_087077d3-f9fa-4d16-aa7c-eb4241cbb293/ovn-controller/0.log" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.845732 4806 generic.go:334] "Generic (PLEG): container finished" podID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerID="9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8" exitCode=2 Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.845768 4806 generic.go:334] "Generic (PLEG): container finished" podID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerID="e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546" exitCode=0 Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.845781 4806 generic.go:334] "Generic (PLEG): container finished" podID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerID="24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7" exitCode=0 Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.845792 4806 generic.go:334] "Generic (PLEG): container finished" podID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerID="ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7" exitCode=0 Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.845801 4806 generic.go:334] "Generic (PLEG): container finished" podID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerID="1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c" exitCode=0 Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.845809 4806 generic.go:334] "Generic (PLEG): container finished" podID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerID="026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b" exitCode=0 Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.845818 4806 generic.go:334] "Generic (PLEG): container finished" podID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerID="5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526" exitCode=143 Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.845827 4806 generic.go:334] "Generic (PLEG): container finished" podID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" containerID="dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28" exitCode=143 Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.845856 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerDied","Data":"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.845899 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerDied","Data":"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.845947 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerDied","Data":"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.845966 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerDied","Data":"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.845979 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerDied","Data":"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.845992 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerDied","Data":"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846006 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846022 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846030 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846041 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846049 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846057 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846200 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846209 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846216 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846235 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerDied","Data":"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846250 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846258 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846266 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846274 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846282 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846289 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846296 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846311 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846318 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846325 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846335 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerDied","Data":"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846345 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846353 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846359 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846366 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846373 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846380 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846387 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846394 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846400 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846407 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846416 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" event={"ID":"087077d3-f9fa-4d16-aa7c-eb4241cbb293","Type":"ContainerDied","Data":"0809ad07fd8b567e1c56cc8cea97f243e1f77063024a08be9619287825b8f072"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846427 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846435 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846442 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846449 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846457 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846464 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846470 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846477 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846484 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846490 4806 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a"} Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846024 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ljxsg" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.846080 4806 scope.go:117] "RemoveContainer" containerID="9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.865177 4806 scope.go:117] "RemoveContainer" containerID="7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.910846 4806 scope.go:117] "RemoveContainer" containerID="e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.917152 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ljxsg"] Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.920964 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ljxsg"] Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.934667 4806 scope.go:117] "RemoveContainer" containerID="24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.951882 4806 scope.go:117] "RemoveContainer" containerID="ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.970211 4806 scope.go:117] "RemoveContainer" containerID="1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c" Dec 04 04:06:39 crc kubenswrapper[4806]: I1204 04:06:39.988694 4806 scope.go:117] "RemoveContainer" containerID="026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.007113 4806 scope.go:117] "RemoveContainer" containerID="5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.051192 4806 scope.go:117] "RemoveContainer" containerID="dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.093544 4806 scope.go:117] "RemoveContainer" containerID="413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.123251 4806 scope.go:117] "RemoveContainer" containerID="9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8" Dec 04 04:06:40 crc kubenswrapper[4806]: E1204 04:06:40.124247 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8\": container with ID starting with 9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8 not found: ID does not exist" containerID="9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.124283 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8"} err="failed to get container status \"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8\": rpc error: code = NotFound desc = could not find container \"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8\": container with ID starting with 9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.124313 4806 scope.go:117] "RemoveContainer" containerID="7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d" Dec 04 04:06:40 crc kubenswrapper[4806]: E1204 04:06:40.124613 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\": container with ID starting with 7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d not found: ID does not exist" containerID="7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.124638 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d"} err="failed to get container status \"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\": rpc error: code = NotFound desc = could not find container \"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\": container with ID starting with 7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.124655 4806 scope.go:117] "RemoveContainer" containerID="e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546" Dec 04 04:06:40 crc kubenswrapper[4806]: E1204 04:06:40.125384 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\": container with ID starting with e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546 not found: ID does not exist" containerID="e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.125419 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546"} err="failed to get container status \"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\": rpc error: code = NotFound desc = could not find container \"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\": container with ID starting with e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.125434 4806 scope.go:117] "RemoveContainer" containerID="24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7" Dec 04 04:06:40 crc kubenswrapper[4806]: E1204 04:06:40.125857 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\": container with ID starting with 24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7 not found: ID does not exist" containerID="24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.125877 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7"} err="failed to get container status \"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\": rpc error: code = NotFound desc = could not find container \"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\": container with ID starting with 24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.125893 4806 scope.go:117] "RemoveContainer" containerID="ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7" Dec 04 04:06:40 crc kubenswrapper[4806]: E1204 04:06:40.126186 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\": container with ID starting with ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7 not found: ID does not exist" containerID="ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.126204 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7"} err="failed to get container status \"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\": rpc error: code = NotFound desc = could not find container \"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\": container with ID starting with ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.126217 4806 scope.go:117] "RemoveContainer" containerID="1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c" Dec 04 04:06:40 crc kubenswrapper[4806]: E1204 04:06:40.126491 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\": container with ID starting with 1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c not found: ID does not exist" containerID="1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.126509 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c"} err="failed to get container status \"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\": rpc error: code = NotFound desc = could not find container \"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\": container with ID starting with 1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.126520 4806 scope.go:117] "RemoveContainer" containerID="026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b" Dec 04 04:06:40 crc kubenswrapper[4806]: E1204 04:06:40.126744 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\": container with ID starting with 026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b not found: ID does not exist" containerID="026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.126761 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b"} err="failed to get container status \"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\": rpc error: code = NotFound desc = could not find container \"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\": container with ID starting with 026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.126771 4806 scope.go:117] "RemoveContainer" containerID="5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526" Dec 04 04:06:40 crc kubenswrapper[4806]: E1204 04:06:40.127082 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\": container with ID starting with 5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526 not found: ID does not exist" containerID="5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.127101 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526"} err="failed to get container status \"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\": rpc error: code = NotFound desc = could not find container \"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\": container with ID starting with 5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.127112 4806 scope.go:117] "RemoveContainer" containerID="dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28" Dec 04 04:06:40 crc kubenswrapper[4806]: E1204 04:06:40.127341 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\": container with ID starting with dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28 not found: ID does not exist" containerID="dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.127358 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28"} err="failed to get container status \"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\": rpc error: code = NotFound desc = could not find container \"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\": container with ID starting with dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.127371 4806 scope.go:117] "RemoveContainer" containerID="413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a" Dec 04 04:06:40 crc kubenswrapper[4806]: E1204 04:06:40.127668 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\": container with ID starting with 413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a not found: ID does not exist" containerID="413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.127685 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a"} err="failed to get container status \"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\": rpc error: code = NotFound desc = could not find container \"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\": container with ID starting with 413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.127697 4806 scope.go:117] "RemoveContainer" containerID="9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.127863 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8"} err="failed to get container status \"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8\": rpc error: code = NotFound desc = could not find container \"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8\": container with ID starting with 9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.127880 4806 scope.go:117] "RemoveContainer" containerID="7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.128091 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d"} err="failed to get container status \"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\": rpc error: code = NotFound desc = could not find container \"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\": container with ID starting with 7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.128106 4806 scope.go:117] "RemoveContainer" containerID="e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.128346 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546"} err="failed to get container status \"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\": rpc error: code = NotFound desc = could not find container \"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\": container with ID starting with e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.128361 4806 scope.go:117] "RemoveContainer" containerID="24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.128564 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7"} err="failed to get container status \"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\": rpc error: code = NotFound desc = could not find container \"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\": container with ID starting with 24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.128578 4806 scope.go:117] "RemoveContainer" containerID="ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.128863 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7"} err="failed to get container status \"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\": rpc error: code = NotFound desc = could not find container \"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\": container with ID starting with ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.128880 4806 scope.go:117] "RemoveContainer" containerID="1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.129111 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c"} err="failed to get container status \"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\": rpc error: code = NotFound desc = could not find container \"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\": container with ID starting with 1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.129129 4806 scope.go:117] "RemoveContainer" containerID="026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.129329 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b"} err="failed to get container status \"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\": rpc error: code = NotFound desc = could not find container \"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\": container with ID starting with 026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.129344 4806 scope.go:117] "RemoveContainer" containerID="5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.129601 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526"} err="failed to get container status \"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\": rpc error: code = NotFound desc = could not find container \"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\": container with ID starting with 5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.129620 4806 scope.go:117] "RemoveContainer" containerID="dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.129824 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28"} err="failed to get container status \"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\": rpc error: code = NotFound desc = could not find container \"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\": container with ID starting with dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.129841 4806 scope.go:117] "RemoveContainer" containerID="413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.130192 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a"} err="failed to get container status \"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\": rpc error: code = NotFound desc = could not find container \"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\": container with ID starting with 413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.130206 4806 scope.go:117] "RemoveContainer" containerID="9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.130533 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8"} err="failed to get container status \"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8\": rpc error: code = NotFound desc = could not find container \"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8\": container with ID starting with 9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.130549 4806 scope.go:117] "RemoveContainer" containerID="7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.130826 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d"} err="failed to get container status \"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\": rpc error: code = NotFound desc = could not find container \"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\": container with ID starting with 7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.130843 4806 scope.go:117] "RemoveContainer" containerID="e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.131182 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546"} err="failed to get container status \"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\": rpc error: code = NotFound desc = could not find container \"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\": container with ID starting with e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.131198 4806 scope.go:117] "RemoveContainer" containerID="24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.131429 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7"} err="failed to get container status \"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\": rpc error: code = NotFound desc = could not find container \"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\": container with ID starting with 24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.131444 4806 scope.go:117] "RemoveContainer" containerID="ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.131704 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7"} err="failed to get container status \"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\": rpc error: code = NotFound desc = could not find container \"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\": container with ID starting with ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.131718 4806 scope.go:117] "RemoveContainer" containerID="1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.132199 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c"} err="failed to get container status \"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\": rpc error: code = NotFound desc = could not find container \"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\": container with ID starting with 1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.132237 4806 scope.go:117] "RemoveContainer" containerID="026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.132572 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b"} err="failed to get container status \"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\": rpc error: code = NotFound desc = could not find container \"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\": container with ID starting with 026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.132596 4806 scope.go:117] "RemoveContainer" containerID="5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.132859 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526"} err="failed to get container status \"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\": rpc error: code = NotFound desc = could not find container \"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\": container with ID starting with 5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.132875 4806 scope.go:117] "RemoveContainer" containerID="dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.133281 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28"} err="failed to get container status \"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\": rpc error: code = NotFound desc = could not find container \"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\": container with ID starting with dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.133302 4806 scope.go:117] "RemoveContainer" containerID="413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.133768 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a"} err="failed to get container status \"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\": rpc error: code = NotFound desc = could not find container \"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\": container with ID starting with 413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.133790 4806 scope.go:117] "RemoveContainer" containerID="9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.134108 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8"} err="failed to get container status \"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8\": rpc error: code = NotFound desc = could not find container \"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8\": container with ID starting with 9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.134132 4806 scope.go:117] "RemoveContainer" containerID="7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.134430 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d"} err="failed to get container status \"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\": rpc error: code = NotFound desc = could not find container \"7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d\": container with ID starting with 7c4d52970b43adb916c7659a199907c3b07a25f172b47313f94d335933522d8d not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.134449 4806 scope.go:117] "RemoveContainer" containerID="e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.134667 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546"} err="failed to get container status \"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\": rpc error: code = NotFound desc = could not find container \"e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546\": container with ID starting with e12bae532ee198197d635ba0a8b8ab321f52b7ead2aeadf69cb0ada8799b8546 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.134689 4806 scope.go:117] "RemoveContainer" containerID="24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.134975 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7"} err="failed to get container status \"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\": rpc error: code = NotFound desc = could not find container \"24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7\": container with ID starting with 24e0bbc2d53dca3cd5c3f9470db1e463c9ecfc3996d75b7e1c840e028d0282d7 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.134996 4806 scope.go:117] "RemoveContainer" containerID="ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.135177 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7"} err="failed to get container status \"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\": rpc error: code = NotFound desc = could not find container \"ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7\": container with ID starting with ee05c629e34b18acc4e02ffa494ee3c6b106ebbefd865bd3aaf8af1b2787bae7 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.135189 4806 scope.go:117] "RemoveContainer" containerID="1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.135453 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c"} err="failed to get container status \"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\": rpc error: code = NotFound desc = could not find container \"1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c\": container with ID starting with 1c492eb7db1c3eb5c08a9df5961efd319cd1f739ee71102988bac8cf8c6d2a9c not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.135472 4806 scope.go:117] "RemoveContainer" containerID="026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.135792 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b"} err="failed to get container status \"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\": rpc error: code = NotFound desc = could not find container \"026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b\": container with ID starting with 026dada4bc709a8a7c2862125f6064edfffb8b2f84e51283eef56ee4246c5e8b not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.135811 4806 scope.go:117] "RemoveContainer" containerID="5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.136787 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526"} err="failed to get container status \"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\": rpc error: code = NotFound desc = could not find container \"5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526\": container with ID starting with 5b3fb2fa3d9a202afd01711f608f1f5035b2283b5b70c59f7f30f3456f4a6526 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.136804 4806 scope.go:117] "RemoveContainer" containerID="dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.137104 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28"} err="failed to get container status \"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\": rpc error: code = NotFound desc = could not find container \"dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28\": container with ID starting with dd12859dbc27613beac9b3fbe2fa3d37b0bdc138164a7aa63a81b760f86a5e28 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.137121 4806 scope.go:117] "RemoveContainer" containerID="413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.137485 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a"} err="failed to get container status \"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\": rpc error: code = NotFound desc = could not find container \"413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a\": container with ID starting with 413923b6c965090f25cf1054869df93cc3d1dcb90af6b09ceea7618ae038e45a not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.137501 4806 scope.go:117] "RemoveContainer" containerID="9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.137672 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8"} err="failed to get container status \"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8\": rpc error: code = NotFound desc = could not find container \"9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8\": container with ID starting with 9fe266ca5d39faf29832667f866e7626827b7c4f57ec0ddbeb30d737ad3da1b8 not found: ID does not exist" Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.855969 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" event={"ID":"b18e6557-daa1-469e-9042-7fcc2d2d5140","Type":"ContainerStarted","Data":"d31ac264bf486f7cf0e752f17e371fe71b281213dc3b34615b77cfe319ea7ee9"} Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.856032 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" event={"ID":"b18e6557-daa1-469e-9042-7fcc2d2d5140","Type":"ContainerStarted","Data":"f25657bc9327d85ddb16e8a2bec7318519656153341ae8f69e441f41f46ea81a"} Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.856043 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" event={"ID":"b18e6557-daa1-469e-9042-7fcc2d2d5140","Type":"ContainerStarted","Data":"232ef331270ac03728b21f26b9195811a075e7b1299115f39776428a2d8b0cf3"} Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.856055 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" event={"ID":"b18e6557-daa1-469e-9042-7fcc2d2d5140","Type":"ContainerStarted","Data":"35e6dd817b143495211dc31be6d399b9aebff4599c49e5e415505d573877b672"} Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.856064 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" event={"ID":"b18e6557-daa1-469e-9042-7fcc2d2d5140","Type":"ContainerStarted","Data":"aeebd459cc8ff7beff1a870473aff3977ffb56de7cbaeb8ba62c2acc45595f2e"} Dec 04 04:06:40 crc kubenswrapper[4806]: I1204 04:06:40.856074 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" event={"ID":"b18e6557-daa1-469e-9042-7fcc2d2d5140","Type":"ContainerStarted","Data":"8328521b5b55f32f3b9806bc0111cc9675442427af669fe7f9fa3c1819fea6e6"} Dec 04 04:06:41 crc kubenswrapper[4806]: I1204 04:06:41.436375 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="087077d3-f9fa-4d16-aa7c-eb4241cbb293" path="/var/lib/kubelet/pods/087077d3-f9fa-4d16-aa7c-eb4241cbb293/volumes" Dec 04 04:06:42 crc kubenswrapper[4806]: I1204 04:06:42.880195 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" event={"ID":"b18e6557-daa1-469e-9042-7fcc2d2d5140","Type":"ContainerStarted","Data":"4ba607b8c22556a0677d7afc4218b59133f03a32f2af96554f50473208818f47"} Dec 04 04:06:45 crc kubenswrapper[4806]: I1204 04:06:45.715566 4806 scope.go:117] "RemoveContainer" containerID="bfb42a576a40b0d1c0a9e0d7fcf9a6c59a3320d5c63147a1f733059c12fcfcd4" Dec 04 04:06:45 crc kubenswrapper[4806]: I1204 04:06:45.904216 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" event={"ID":"b18e6557-daa1-469e-9042-7fcc2d2d5140","Type":"ContainerStarted","Data":"700e0f90c33cdcd7212c035daca2afbe7b83e1b38bd1634ede02e0e4f7d9b538"} Dec 04 04:06:45 crc kubenswrapper[4806]: I1204 04:06:45.904576 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:45 crc kubenswrapper[4806]: I1204 04:06:45.971118 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:45 crc kubenswrapper[4806]: I1204 04:06:45.991108 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" podStartSLOduration=6.991075716 podStartE2EDuration="6.991075716s" podCreationTimestamp="2025-12-04 04:06:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:06:45.985713486 +0000 UTC m=+720.844226434" watchObservedRunningTime="2025-12-04 04:06:45.991075716 +0000 UTC m=+720.849588644" Dec 04 04:06:46 crc kubenswrapper[4806]: I1204 04:06:46.912776 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:46 crc kubenswrapper[4806]: I1204 04:06:46.913458 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:46 crc kubenswrapper[4806]: I1204 04:06:46.949757 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:06:49 crc kubenswrapper[4806]: I1204 04:06:49.424623 4806 scope.go:117] "RemoveContainer" containerID="d6efe030cffcc9ae048a86abeffad5784f713909803cbabe74db70e9e5212015" Dec 04 04:06:49 crc kubenswrapper[4806]: E1204 04:06:49.425452 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-dsclk_openshift-multus(4f13bcfc-c287-40fd-b8f6-e4bbea8c6577)\"" pod="openshift-multus/multus-dsclk" podUID="4f13bcfc-c287-40fd-b8f6-e4bbea8c6577" Dec 04 04:06:57 crc kubenswrapper[4806]: I1204 04:06:57.048059 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:06:57 crc kubenswrapper[4806]: I1204 04:06:57.048907 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:07:03 crc kubenswrapper[4806]: I1204 04:07:03.424207 4806 scope.go:117] "RemoveContainer" containerID="d6efe030cffcc9ae048a86abeffad5784f713909803cbabe74db70e9e5212015" Dec 04 04:07:04 crc kubenswrapper[4806]: I1204 04:07:04.033959 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-dsclk_4f13bcfc-c287-40fd-b8f6-e4bbea8c6577/kube-multus/2.log" Dec 04 04:07:04 crc kubenswrapper[4806]: I1204 04:07:04.034515 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-dsclk" event={"ID":"4f13bcfc-c287-40fd-b8f6-e4bbea8c6577","Type":"ContainerStarted","Data":"0f333dd1667e0ae06d9833f630a452b77f0738d14b8f00c307ec0c312616c8a4"} Dec 04 04:07:09 crc kubenswrapper[4806]: I1204 04:07:09.482566 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5cqmh" Dec 04 04:07:10 crc kubenswrapper[4806]: I1204 04:07:10.220283 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b"] Dec 04 04:07:10 crc kubenswrapper[4806]: I1204 04:07:10.221853 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" Dec 04 04:07:10 crc kubenswrapper[4806]: I1204 04:07:10.225045 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 04:07:10 crc kubenswrapper[4806]: I1204 04:07:10.239190 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b"] Dec 04 04:07:10 crc kubenswrapper[4806]: I1204 04:07:10.338917 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8981d20c-af12-4497-abc8-f17930ec12c1-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b\" (UID: \"8981d20c-af12-4497-abc8-f17930ec12c1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" Dec 04 04:07:10 crc kubenswrapper[4806]: I1204 04:07:10.339030 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkkv9\" (UniqueName: \"kubernetes.io/projected/8981d20c-af12-4497-abc8-f17930ec12c1-kube-api-access-rkkv9\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b\" (UID: \"8981d20c-af12-4497-abc8-f17930ec12c1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" Dec 04 04:07:10 crc kubenswrapper[4806]: I1204 04:07:10.339091 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8981d20c-af12-4497-abc8-f17930ec12c1-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b\" (UID: \"8981d20c-af12-4497-abc8-f17930ec12c1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" Dec 04 04:07:10 crc kubenswrapper[4806]: I1204 04:07:10.441579 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8981d20c-af12-4497-abc8-f17930ec12c1-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b\" (UID: \"8981d20c-af12-4497-abc8-f17930ec12c1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" Dec 04 04:07:10 crc kubenswrapper[4806]: I1204 04:07:10.441873 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8981d20c-af12-4497-abc8-f17930ec12c1-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b\" (UID: \"8981d20c-af12-4497-abc8-f17930ec12c1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" Dec 04 04:07:10 crc kubenswrapper[4806]: I1204 04:07:10.441900 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkkv9\" (UniqueName: \"kubernetes.io/projected/8981d20c-af12-4497-abc8-f17930ec12c1-kube-api-access-rkkv9\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b\" (UID: \"8981d20c-af12-4497-abc8-f17930ec12c1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" Dec 04 04:07:10 crc kubenswrapper[4806]: I1204 04:07:10.442677 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8981d20c-af12-4497-abc8-f17930ec12c1-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b\" (UID: \"8981d20c-af12-4497-abc8-f17930ec12c1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" Dec 04 04:07:10 crc kubenswrapper[4806]: I1204 04:07:10.443024 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8981d20c-af12-4497-abc8-f17930ec12c1-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b\" (UID: \"8981d20c-af12-4497-abc8-f17930ec12c1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" Dec 04 04:07:10 crc kubenswrapper[4806]: I1204 04:07:10.467872 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkkv9\" (UniqueName: \"kubernetes.io/projected/8981d20c-af12-4497-abc8-f17930ec12c1-kube-api-access-rkkv9\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b\" (UID: \"8981d20c-af12-4497-abc8-f17930ec12c1\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" Dec 04 04:07:10 crc kubenswrapper[4806]: I1204 04:07:10.542424 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" Dec 04 04:07:10 crc kubenswrapper[4806]: I1204 04:07:10.961873 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b"] Dec 04 04:07:11 crc kubenswrapper[4806]: I1204 04:07:11.088716 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" event={"ID":"8981d20c-af12-4497-abc8-f17930ec12c1","Type":"ContainerStarted","Data":"9bc048d5776f29a1a230c0cb9a064324833a58960207d14c2949a1154971c890"} Dec 04 04:07:12 crc kubenswrapper[4806]: I1204 04:07:12.097854 4806 generic.go:334] "Generic (PLEG): container finished" podID="8981d20c-af12-4497-abc8-f17930ec12c1" containerID="2eef85c65333b0bdc8bc69dabbbe700e8e095ecb1c62c92baa10f2af38035200" exitCode=0 Dec 04 04:07:12 crc kubenswrapper[4806]: I1204 04:07:12.097961 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" event={"ID":"8981d20c-af12-4497-abc8-f17930ec12c1","Type":"ContainerDied","Data":"2eef85c65333b0bdc8bc69dabbbe700e8e095ecb1c62c92baa10f2af38035200"} Dec 04 04:07:14 crc kubenswrapper[4806]: I1204 04:07:14.119481 4806 generic.go:334] "Generic (PLEG): container finished" podID="8981d20c-af12-4497-abc8-f17930ec12c1" containerID="df2197ebd8351ac1fa81917df46e10e0f5deba055990cb0bfc86729f240d5887" exitCode=0 Dec 04 04:07:14 crc kubenswrapper[4806]: I1204 04:07:14.119588 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" event={"ID":"8981d20c-af12-4497-abc8-f17930ec12c1","Type":"ContainerDied","Data":"df2197ebd8351ac1fa81917df46e10e0f5deba055990cb0bfc86729f240d5887"} Dec 04 04:07:15 crc kubenswrapper[4806]: I1204 04:07:15.130114 4806 generic.go:334] "Generic (PLEG): container finished" podID="8981d20c-af12-4497-abc8-f17930ec12c1" containerID="23507618fdbd6ebd13614a862f6d1393dc45b4ff222c6bc28cfcd4b7624e0ba3" exitCode=0 Dec 04 04:07:15 crc kubenswrapper[4806]: I1204 04:07:15.130665 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" event={"ID":"8981d20c-af12-4497-abc8-f17930ec12c1","Type":"ContainerDied","Data":"23507618fdbd6ebd13614a862f6d1393dc45b4ff222c6bc28cfcd4b7624e0ba3"} Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.139732 4806 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.457596 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.551470 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8981d20c-af12-4497-abc8-f17930ec12c1-bundle\") pod \"8981d20c-af12-4497-abc8-f17930ec12c1\" (UID: \"8981d20c-af12-4497-abc8-f17930ec12c1\") " Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.551569 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkkv9\" (UniqueName: \"kubernetes.io/projected/8981d20c-af12-4497-abc8-f17930ec12c1-kube-api-access-rkkv9\") pod \"8981d20c-af12-4497-abc8-f17930ec12c1\" (UID: \"8981d20c-af12-4497-abc8-f17930ec12c1\") " Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.551715 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8981d20c-af12-4497-abc8-f17930ec12c1-util\") pod \"8981d20c-af12-4497-abc8-f17930ec12c1\" (UID: \"8981d20c-af12-4497-abc8-f17930ec12c1\") " Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.552499 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8981d20c-af12-4497-abc8-f17930ec12c1-bundle" (OuterVolumeSpecName: "bundle") pod "8981d20c-af12-4497-abc8-f17930ec12c1" (UID: "8981d20c-af12-4497-abc8-f17930ec12c1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.558146 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8981d20c-af12-4497-abc8-f17930ec12c1-kube-api-access-rkkv9" (OuterVolumeSpecName: "kube-api-access-rkkv9") pod "8981d20c-af12-4497-abc8-f17930ec12c1" (UID: "8981d20c-af12-4497-abc8-f17930ec12c1"). InnerVolumeSpecName "kube-api-access-rkkv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.567800 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8981d20c-af12-4497-abc8-f17930ec12c1-util" (OuterVolumeSpecName: "util") pod "8981d20c-af12-4497-abc8-f17930ec12c1" (UID: "8981d20c-af12-4497-abc8-f17930ec12c1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.653063 4806 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8981d20c-af12-4497-abc8-f17930ec12c1-util\") on node \"crc\" DevicePath \"\"" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.653122 4806 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8981d20c-af12-4497-abc8-f17930ec12c1-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.653134 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkkv9\" (UniqueName: \"kubernetes.io/projected/8981d20c-af12-4497-abc8-f17930ec12c1-kube-api-access-rkkv9\") on node \"crc\" DevicePath \"\"" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.745683 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pwlns"] Dec 04 04:07:16 crc kubenswrapper[4806]: E1204 04:07:16.746086 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8981d20c-af12-4497-abc8-f17930ec12c1" containerName="pull" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.746105 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="8981d20c-af12-4497-abc8-f17930ec12c1" containerName="pull" Dec 04 04:07:16 crc kubenswrapper[4806]: E1204 04:07:16.746122 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8981d20c-af12-4497-abc8-f17930ec12c1" containerName="extract" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.746130 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="8981d20c-af12-4497-abc8-f17930ec12c1" containerName="extract" Dec 04 04:07:16 crc kubenswrapper[4806]: E1204 04:07:16.746153 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8981d20c-af12-4497-abc8-f17930ec12c1" containerName="util" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.746164 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="8981d20c-af12-4497-abc8-f17930ec12c1" containerName="util" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.746299 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="8981d20c-af12-4497-abc8-f17930ec12c1" containerName="extract" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.747383 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.767789 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pwlns"] Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.855582 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6df24743-9526-4457-9f48-f8639a96977f-catalog-content\") pod \"redhat-operators-pwlns\" (UID: \"6df24743-9526-4457-9f48-f8639a96977f\") " pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.855729 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsbx6\" (UniqueName: \"kubernetes.io/projected/6df24743-9526-4457-9f48-f8639a96977f-kube-api-access-wsbx6\") pod \"redhat-operators-pwlns\" (UID: \"6df24743-9526-4457-9f48-f8639a96977f\") " pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.855785 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6df24743-9526-4457-9f48-f8639a96977f-utilities\") pod \"redhat-operators-pwlns\" (UID: \"6df24743-9526-4457-9f48-f8639a96977f\") " pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.957074 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6df24743-9526-4457-9f48-f8639a96977f-utilities\") pod \"redhat-operators-pwlns\" (UID: \"6df24743-9526-4457-9f48-f8639a96977f\") " pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.957140 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6df24743-9526-4457-9f48-f8639a96977f-catalog-content\") pod \"redhat-operators-pwlns\" (UID: \"6df24743-9526-4457-9f48-f8639a96977f\") " pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.957199 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsbx6\" (UniqueName: \"kubernetes.io/projected/6df24743-9526-4457-9f48-f8639a96977f-kube-api-access-wsbx6\") pod \"redhat-operators-pwlns\" (UID: \"6df24743-9526-4457-9f48-f8639a96977f\") " pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.957851 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6df24743-9526-4457-9f48-f8639a96977f-utilities\") pod \"redhat-operators-pwlns\" (UID: \"6df24743-9526-4457-9f48-f8639a96977f\") " pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.958063 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6df24743-9526-4457-9f48-f8639a96977f-catalog-content\") pod \"redhat-operators-pwlns\" (UID: \"6df24743-9526-4457-9f48-f8639a96977f\") " pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:16 crc kubenswrapper[4806]: I1204 04:07:16.986704 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsbx6\" (UniqueName: \"kubernetes.io/projected/6df24743-9526-4457-9f48-f8639a96977f-kube-api-access-wsbx6\") pod \"redhat-operators-pwlns\" (UID: \"6df24743-9526-4457-9f48-f8639a96977f\") " pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:17 crc kubenswrapper[4806]: I1204 04:07:17.061256 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:17 crc kubenswrapper[4806]: I1204 04:07:17.157485 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" event={"ID":"8981d20c-af12-4497-abc8-f17930ec12c1","Type":"ContainerDied","Data":"9bc048d5776f29a1a230c0cb9a064324833a58960207d14c2949a1154971c890"} Dec 04 04:07:17 crc kubenswrapper[4806]: I1204 04:07:17.158025 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bc048d5776f29a1a230c0cb9a064324833a58960207d14c2949a1154971c890" Dec 04 04:07:17 crc kubenswrapper[4806]: I1204 04:07:17.157602 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b" Dec 04 04:07:17 crc kubenswrapper[4806]: I1204 04:07:17.597853 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pwlns"] Dec 04 04:07:18 crc kubenswrapper[4806]: I1204 04:07:18.169132 4806 generic.go:334] "Generic (PLEG): container finished" podID="6df24743-9526-4457-9f48-f8639a96977f" containerID="fc96cfaa23666ab41da5aac3901e25a3328ec4e7febf98918543fd73ec85649c" exitCode=0 Dec 04 04:07:18 crc kubenswrapper[4806]: I1204 04:07:18.170336 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwlns" event={"ID":"6df24743-9526-4457-9f48-f8639a96977f","Type":"ContainerDied","Data":"fc96cfaa23666ab41da5aac3901e25a3328ec4e7febf98918543fd73ec85649c"} Dec 04 04:07:18 crc kubenswrapper[4806]: I1204 04:07:18.170426 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwlns" event={"ID":"6df24743-9526-4457-9f48-f8639a96977f","Type":"ContainerStarted","Data":"671ddf7e9e30c07a75c79ec10f657d6dde848606b7c9c384392c9ef1a43c5a48"} Dec 04 04:07:19 crc kubenswrapper[4806]: I1204 04:07:19.178600 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwlns" event={"ID":"6df24743-9526-4457-9f48-f8639a96977f","Type":"ContainerStarted","Data":"8f6465e25ee5f249a999c4a24b9d50faf319818d217505569fc04aac34ff6f15"} Dec 04 04:07:20 crc kubenswrapper[4806]: I1204 04:07:20.187887 4806 generic.go:334] "Generic (PLEG): container finished" podID="6df24743-9526-4457-9f48-f8639a96977f" containerID="8f6465e25ee5f249a999c4a24b9d50faf319818d217505569fc04aac34ff6f15" exitCode=0 Dec 04 04:07:20 crc kubenswrapper[4806]: I1204 04:07:20.187968 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwlns" event={"ID":"6df24743-9526-4457-9f48-f8639a96977f","Type":"ContainerDied","Data":"8f6465e25ee5f249a999c4a24b9d50faf319818d217505569fc04aac34ff6f15"} Dec 04 04:07:20 crc kubenswrapper[4806]: I1204 04:07:20.805278 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-zz62j"] Dec 04 04:07:20 crc kubenswrapper[4806]: I1204 04:07:20.806988 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-zz62j" Dec 04 04:07:20 crc kubenswrapper[4806]: I1204 04:07:20.810568 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 04 04:07:20 crc kubenswrapper[4806]: I1204 04:07:20.810596 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-gzw4p" Dec 04 04:07:20 crc kubenswrapper[4806]: I1204 04:07:20.810568 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 04 04:07:20 crc kubenswrapper[4806]: I1204 04:07:20.831740 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw6dt\" (UniqueName: \"kubernetes.io/projected/93c65cc3-ea7f-4f77-af33-8acdcd03432c-kube-api-access-xw6dt\") pod \"nmstate-operator-5b5b58f5c8-zz62j\" (UID: \"93c65cc3-ea7f-4f77-af33-8acdcd03432c\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-zz62j" Dec 04 04:07:20 crc kubenswrapper[4806]: I1204 04:07:20.861455 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-zz62j"] Dec 04 04:07:20 crc kubenswrapper[4806]: I1204 04:07:20.933975 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw6dt\" (UniqueName: \"kubernetes.io/projected/93c65cc3-ea7f-4f77-af33-8acdcd03432c-kube-api-access-xw6dt\") pod \"nmstate-operator-5b5b58f5c8-zz62j\" (UID: \"93c65cc3-ea7f-4f77-af33-8acdcd03432c\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-zz62j" Dec 04 04:07:20 crc kubenswrapper[4806]: I1204 04:07:20.960089 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw6dt\" (UniqueName: \"kubernetes.io/projected/93c65cc3-ea7f-4f77-af33-8acdcd03432c-kube-api-access-xw6dt\") pod \"nmstate-operator-5b5b58f5c8-zz62j\" (UID: \"93c65cc3-ea7f-4f77-af33-8acdcd03432c\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-zz62j" Dec 04 04:07:21 crc kubenswrapper[4806]: I1204 04:07:21.124704 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-zz62j" Dec 04 04:07:21 crc kubenswrapper[4806]: I1204 04:07:21.210347 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwlns" event={"ID":"6df24743-9526-4457-9f48-f8639a96977f","Type":"ContainerStarted","Data":"f0ef149b28b6df5facad56f98fae2579aaf63d57cd97e0a096c849b1b054c1d5"} Dec 04 04:07:21 crc kubenswrapper[4806]: I1204 04:07:21.246612 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pwlns" podStartSLOduration=2.852097622 podStartE2EDuration="5.24657923s" podCreationTimestamp="2025-12-04 04:07:16 +0000 UTC" firstStartedPulling="2025-12-04 04:07:18.172496606 +0000 UTC m=+753.031009534" lastFinishedPulling="2025-12-04 04:07:20.566978214 +0000 UTC m=+755.425491142" observedRunningTime="2025-12-04 04:07:21.240873942 +0000 UTC m=+756.099386890" watchObservedRunningTime="2025-12-04 04:07:21.24657923 +0000 UTC m=+756.105092168" Dec 04 04:07:21 crc kubenswrapper[4806]: I1204 04:07:21.386165 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-zz62j"] Dec 04 04:07:22 crc kubenswrapper[4806]: I1204 04:07:22.217419 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-zz62j" event={"ID":"93c65cc3-ea7f-4f77-af33-8acdcd03432c","Type":"ContainerStarted","Data":"8fc139eab7ce9442bf9d4fd085e62a0926c04e9d2ff3e24440b706f80368d728"} Dec 04 04:07:24 crc kubenswrapper[4806]: I1204 04:07:24.236439 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-zz62j" event={"ID":"93c65cc3-ea7f-4f77-af33-8acdcd03432c","Type":"ContainerStarted","Data":"e27bed2738b37f4372fd790c806f1847a1f45b870e4155a4bad15fc18eb91c51"} Dec 04 04:07:24 crc kubenswrapper[4806]: I1204 04:07:24.263257 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-zz62j" podStartSLOduration=2.053701426 podStartE2EDuration="4.263224988s" podCreationTimestamp="2025-12-04 04:07:20 +0000 UTC" firstStartedPulling="2025-12-04 04:07:21.395680422 +0000 UTC m=+756.254193350" lastFinishedPulling="2025-12-04 04:07:23.605203984 +0000 UTC m=+758.463716912" observedRunningTime="2025-12-04 04:07:24.257033187 +0000 UTC m=+759.115546115" watchObservedRunningTime="2025-12-04 04:07:24.263224988 +0000 UTC m=+759.121737916" Dec 04 04:07:27 crc kubenswrapper[4806]: I1204 04:07:27.048008 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:07:27 crc kubenswrapper[4806]: I1204 04:07:27.049150 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:07:27 crc kubenswrapper[4806]: I1204 04:07:27.049254 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 04:07:27 crc kubenswrapper[4806]: I1204 04:07:27.050226 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f0bc0378fad2cb0fb66c2e5ff86d10f9a724302404985dba5e351fcfd4397778"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 04:07:27 crc kubenswrapper[4806]: I1204 04:07:27.050309 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://f0bc0378fad2cb0fb66c2e5ff86d10f9a724302404985dba5e351fcfd4397778" gracePeriod=600 Dec 04 04:07:27 crc kubenswrapper[4806]: I1204 04:07:27.061989 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:27 crc kubenswrapper[4806]: I1204 04:07:27.062054 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:27 crc kubenswrapper[4806]: I1204 04:07:27.107276 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:27 crc kubenswrapper[4806]: I1204 04:07:27.294989 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:28 crc kubenswrapper[4806]: I1204 04:07:28.265705 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="f0bc0378fad2cb0fb66c2e5ff86d10f9a724302404985dba5e351fcfd4397778" exitCode=0 Dec 04 04:07:28 crc kubenswrapper[4806]: I1204 04:07:28.266039 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"f0bc0378fad2cb0fb66c2e5ff86d10f9a724302404985dba5e351fcfd4397778"} Dec 04 04:07:28 crc kubenswrapper[4806]: I1204 04:07:28.267597 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"4ce73f616eb3f88f0847c6436643d27fc2337de01de1b065778733e8ceef4522"} Dec 04 04:07:28 crc kubenswrapper[4806]: I1204 04:07:28.267633 4806 scope.go:117] "RemoveContainer" containerID="14fd61c0a81ae92a7b878dd7a4719064b58b4a7d3bf90d5f039134a7e74c54ac" Dec 04 04:07:29 crc kubenswrapper[4806]: I1204 04:07:29.533144 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pwlns"] Dec 04 04:07:29 crc kubenswrapper[4806]: I1204 04:07:29.533871 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pwlns" podUID="6df24743-9526-4457-9f48-f8639a96977f" containerName="registry-server" containerID="cri-o://f0ef149b28b6df5facad56f98fae2579aaf63d57cd97e0a096c849b1b054c1d5" gracePeriod=2 Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.397351 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-dlkzr"] Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.399197 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-dlkzr" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.402075 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-zv2sn" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.405783 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnzkn\" (UniqueName: \"kubernetes.io/projected/7b294b89-f409-40b7-8ba4-72e53da24a6f-kube-api-access-dnzkn\") pod \"nmstate-metrics-7f946cbc9-dlkzr\" (UID: \"7b294b89-f409-40b7-8ba4-72e53da24a6f\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-dlkzr" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.427161 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68"] Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.428455 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.432808 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-dlkzr"] Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.449005 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-pdg7x"] Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.450183 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-pdg7x" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.451640 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.474845 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68"] Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.506918 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnzkn\" (UniqueName: \"kubernetes.io/projected/7b294b89-f409-40b7-8ba4-72e53da24a6f-kube-api-access-dnzkn\") pod \"nmstate-metrics-7f946cbc9-dlkzr\" (UID: \"7b294b89-f409-40b7-8ba4-72e53da24a6f\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-dlkzr" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.552373 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnzkn\" (UniqueName: \"kubernetes.io/projected/7b294b89-f409-40b7-8ba4-72e53da24a6f-kube-api-access-dnzkn\") pod \"nmstate-metrics-7f946cbc9-dlkzr\" (UID: \"7b294b89-f409-40b7-8ba4-72e53da24a6f\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-dlkzr" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.608891 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/843789a7-01bb-42c0-9f30-aa52af788147-dbus-socket\") pod \"nmstate-handler-pdg7x\" (UID: \"843789a7-01bb-42c0-9f30-aa52af788147\") " pod="openshift-nmstate/nmstate-handler-pdg7x" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.609356 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/843789a7-01bb-42c0-9f30-aa52af788147-ovs-socket\") pod \"nmstate-handler-pdg7x\" (UID: \"843789a7-01bb-42c0-9f30-aa52af788147\") " pod="openshift-nmstate/nmstate-handler-pdg7x" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.609380 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/843789a7-01bb-42c0-9f30-aa52af788147-nmstate-lock\") pod \"nmstate-handler-pdg7x\" (UID: \"843789a7-01bb-42c0-9f30-aa52af788147\") " pod="openshift-nmstate/nmstate-handler-pdg7x" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.609412 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mw7r\" (UniqueName: \"kubernetes.io/projected/eae97ef2-3aeb-4303-b26f-0f84a75abda0-kube-api-access-9mw7r\") pod \"nmstate-webhook-5f6d4c5ccb-q9x68\" (UID: \"eae97ef2-3aeb-4303-b26f-0f84a75abda0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.609469 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc9mq\" (UniqueName: \"kubernetes.io/projected/843789a7-01bb-42c0-9f30-aa52af788147-kube-api-access-sc9mq\") pod \"nmstate-handler-pdg7x\" (UID: \"843789a7-01bb-42c0-9f30-aa52af788147\") " pod="openshift-nmstate/nmstate-handler-pdg7x" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.609490 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/eae97ef2-3aeb-4303-b26f-0f84a75abda0-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-q9x68\" (UID: \"eae97ef2-3aeb-4303-b26f-0f84a75abda0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.645048 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx"] Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.646113 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.648747 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.649860 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-cl5gc" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.649914 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.661524 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx"] Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.711159 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/843789a7-01bb-42c0-9f30-aa52af788147-dbus-socket\") pod \"nmstate-handler-pdg7x\" (UID: \"843789a7-01bb-42c0-9f30-aa52af788147\") " pod="openshift-nmstate/nmstate-handler-pdg7x" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.711220 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/843789a7-01bb-42c0-9f30-aa52af788147-ovs-socket\") pod \"nmstate-handler-pdg7x\" (UID: \"843789a7-01bb-42c0-9f30-aa52af788147\") " pod="openshift-nmstate/nmstate-handler-pdg7x" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.711247 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/843789a7-01bb-42c0-9f30-aa52af788147-nmstate-lock\") pod \"nmstate-handler-pdg7x\" (UID: \"843789a7-01bb-42c0-9f30-aa52af788147\") " pod="openshift-nmstate/nmstate-handler-pdg7x" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.711279 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mw7r\" (UniqueName: \"kubernetes.io/projected/eae97ef2-3aeb-4303-b26f-0f84a75abda0-kube-api-access-9mw7r\") pod \"nmstate-webhook-5f6d4c5ccb-q9x68\" (UID: \"eae97ef2-3aeb-4303-b26f-0f84a75abda0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.711335 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc9mq\" (UniqueName: \"kubernetes.io/projected/843789a7-01bb-42c0-9f30-aa52af788147-kube-api-access-sc9mq\") pod \"nmstate-handler-pdg7x\" (UID: \"843789a7-01bb-42c0-9f30-aa52af788147\") " pod="openshift-nmstate/nmstate-handler-pdg7x" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.711359 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/eae97ef2-3aeb-4303-b26f-0f84a75abda0-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-q9x68\" (UID: \"eae97ef2-3aeb-4303-b26f-0f84a75abda0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68" Dec 04 04:07:30 crc kubenswrapper[4806]: E1204 04:07:30.711541 4806 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 04 04:07:30 crc kubenswrapper[4806]: E1204 04:07:30.711614 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eae97ef2-3aeb-4303-b26f-0f84a75abda0-tls-key-pair podName:eae97ef2-3aeb-4303-b26f-0f84a75abda0 nodeName:}" failed. No retries permitted until 2025-12-04 04:07:31.211588952 +0000 UTC m=+766.070101880 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/eae97ef2-3aeb-4303-b26f-0f84a75abda0-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-q9x68" (UID: "eae97ef2-3aeb-4303-b26f-0f84a75abda0") : secret "openshift-nmstate-webhook" not found Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.711671 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/843789a7-01bb-42c0-9f30-aa52af788147-dbus-socket\") pod \"nmstate-handler-pdg7x\" (UID: \"843789a7-01bb-42c0-9f30-aa52af788147\") " pod="openshift-nmstate/nmstate-handler-pdg7x" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.712051 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/843789a7-01bb-42c0-9f30-aa52af788147-ovs-socket\") pod \"nmstate-handler-pdg7x\" (UID: \"843789a7-01bb-42c0-9f30-aa52af788147\") " pod="openshift-nmstate/nmstate-handler-pdg7x" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.712096 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/843789a7-01bb-42c0-9f30-aa52af788147-nmstate-lock\") pod \"nmstate-handler-pdg7x\" (UID: \"843789a7-01bb-42c0-9f30-aa52af788147\") " pod="openshift-nmstate/nmstate-handler-pdg7x" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.718299 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-dlkzr" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.729791 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc9mq\" (UniqueName: \"kubernetes.io/projected/843789a7-01bb-42c0-9f30-aa52af788147-kube-api-access-sc9mq\") pod \"nmstate-handler-pdg7x\" (UID: \"843789a7-01bb-42c0-9f30-aa52af788147\") " pod="openshift-nmstate/nmstate-handler-pdg7x" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.730052 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mw7r\" (UniqueName: \"kubernetes.io/projected/eae97ef2-3aeb-4303-b26f-0f84a75abda0-kube-api-access-9mw7r\") pod \"nmstate-webhook-5f6d4c5ccb-q9x68\" (UID: \"eae97ef2-3aeb-4303-b26f-0f84a75abda0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.766348 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-pdg7x" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.813404 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b0e181e1-d1fa-460f-963f-cbd235500b2e-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-572xx\" (UID: \"b0e181e1-d1fa-460f-963f-cbd235500b2e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.813511 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0e181e1-d1fa-460f-963f-cbd235500b2e-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-572xx\" (UID: \"b0e181e1-d1fa-460f-963f-cbd235500b2e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.813610 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7527\" (UniqueName: \"kubernetes.io/projected/b0e181e1-d1fa-460f-963f-cbd235500b2e-kube-api-access-c7527\") pod \"nmstate-console-plugin-7fbb5f6569-572xx\" (UID: \"b0e181e1-d1fa-460f-963f-cbd235500b2e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.917813 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7527\" (UniqueName: \"kubernetes.io/projected/b0e181e1-d1fa-460f-963f-cbd235500b2e-kube-api-access-c7527\") pod \"nmstate-console-plugin-7fbb5f6569-572xx\" (UID: \"b0e181e1-d1fa-460f-963f-cbd235500b2e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.918250 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b0e181e1-d1fa-460f-963f-cbd235500b2e-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-572xx\" (UID: \"b0e181e1-d1fa-460f-963f-cbd235500b2e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.918320 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0e181e1-d1fa-460f-963f-cbd235500b2e-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-572xx\" (UID: \"b0e181e1-d1fa-460f-963f-cbd235500b2e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx" Dec 04 04:07:30 crc kubenswrapper[4806]: E1204 04:07:30.918567 4806 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 04 04:07:30 crc kubenswrapper[4806]: E1204 04:07:30.918652 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b0e181e1-d1fa-460f-963f-cbd235500b2e-plugin-serving-cert podName:b0e181e1-d1fa-460f-963f-cbd235500b2e nodeName:}" failed. No retries permitted until 2025-12-04 04:07:31.418628243 +0000 UTC m=+766.277141181 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/b0e181e1-d1fa-460f-963f-cbd235500b2e-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-572xx" (UID: "b0e181e1-d1fa-460f-963f-cbd235500b2e") : secret "plugin-serving-cert" not found Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.919807 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b0e181e1-d1fa-460f-963f-cbd235500b2e-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-572xx\" (UID: \"b0e181e1-d1fa-460f-963f-cbd235500b2e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.961576 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7527\" (UniqueName: \"kubernetes.io/projected/b0e181e1-d1fa-460f-963f-cbd235500b2e-kube-api-access-c7527\") pod \"nmstate-console-plugin-7fbb5f6569-572xx\" (UID: \"b0e181e1-d1fa-460f-963f-cbd235500b2e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx" Dec 04 04:07:30 crc kubenswrapper[4806]: I1204 04:07:30.993120 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7d545d4ccb-h5stl"] Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.001052 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.006752 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7d545d4ccb-h5stl"] Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.024040 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/82cf8319-340a-4d1d-92e4-adfadea8a5cf-console-oauth-config\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.024108 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bgsf\" (UniqueName: \"kubernetes.io/projected/82cf8319-340a-4d1d-92e4-adfadea8a5cf-kube-api-access-2bgsf\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.024146 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82cf8319-340a-4d1d-92e4-adfadea8a5cf-service-ca\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.024177 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82cf8319-340a-4d1d-92e4-adfadea8a5cf-trusted-ca-bundle\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.024245 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/82cf8319-340a-4d1d-92e4-adfadea8a5cf-console-serving-cert\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.024287 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/82cf8319-340a-4d1d-92e4-adfadea8a5cf-oauth-serving-cert\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.024306 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/82cf8319-340a-4d1d-92e4-adfadea8a5cf-console-config\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.133895 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-dlkzr"] Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.135799 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/82cf8319-340a-4d1d-92e4-adfadea8a5cf-console-serving-cert\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.135859 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/82cf8319-340a-4d1d-92e4-adfadea8a5cf-oauth-serving-cert\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.135894 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/82cf8319-340a-4d1d-92e4-adfadea8a5cf-console-config\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.135965 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/82cf8319-340a-4d1d-92e4-adfadea8a5cf-console-oauth-config\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.136012 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bgsf\" (UniqueName: \"kubernetes.io/projected/82cf8319-340a-4d1d-92e4-adfadea8a5cf-kube-api-access-2bgsf\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.136057 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82cf8319-340a-4d1d-92e4-adfadea8a5cf-service-ca\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.136094 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82cf8319-340a-4d1d-92e4-adfadea8a5cf-trusted-ca-bundle\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.139571 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82cf8319-340a-4d1d-92e4-adfadea8a5cf-trusted-ca-bundle\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.141673 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/82cf8319-340a-4d1d-92e4-adfadea8a5cf-oauth-serving-cert\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.141748 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82cf8319-340a-4d1d-92e4-adfadea8a5cf-service-ca\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.143294 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/82cf8319-340a-4d1d-92e4-adfadea8a5cf-console-config\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.147045 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/82cf8319-340a-4d1d-92e4-adfadea8a5cf-console-oauth-config\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.153454 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/82cf8319-340a-4d1d-92e4-adfadea8a5cf-console-serving-cert\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.161489 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bgsf\" (UniqueName: \"kubernetes.io/projected/82cf8319-340a-4d1d-92e4-adfadea8a5cf-kube-api-access-2bgsf\") pod \"console-7d545d4ccb-h5stl\" (UID: \"82cf8319-340a-4d1d-92e4-adfadea8a5cf\") " pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.238857 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/eae97ef2-3aeb-4303-b26f-0f84a75abda0-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-q9x68\" (UID: \"eae97ef2-3aeb-4303-b26f-0f84a75abda0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.255035 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/eae97ef2-3aeb-4303-b26f-0f84a75abda0-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-q9x68\" (UID: \"eae97ef2-3aeb-4303-b26f-0f84a75abda0\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.294304 4806 generic.go:334] "Generic (PLEG): container finished" podID="6df24743-9526-4457-9f48-f8639a96977f" containerID="f0ef149b28b6df5facad56f98fae2579aaf63d57cd97e0a096c849b1b054c1d5" exitCode=0 Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.294372 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwlns" event={"ID":"6df24743-9526-4457-9f48-f8639a96977f","Type":"ContainerDied","Data":"f0ef149b28b6df5facad56f98fae2579aaf63d57cd97e0a096c849b1b054c1d5"} Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.296025 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-pdg7x" event={"ID":"843789a7-01bb-42c0-9f30-aa52af788147","Type":"ContainerStarted","Data":"3d8146bd4c96e17ecafd6c6d14caf894de615315d191641ab24213b152a67427"} Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.297664 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-dlkzr" event={"ID":"7b294b89-f409-40b7-8ba4-72e53da24a6f","Type":"ContainerStarted","Data":"a5c4f5e855523f74164f1a866fcaa9102191267ed6ae3190126545d2c0978a9d"} Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.344258 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.372424 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.450392 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0e181e1-d1fa-460f-963f-cbd235500b2e-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-572xx\" (UID: \"b0e181e1-d1fa-460f-963f-cbd235500b2e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.454738 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0e181e1-d1fa-460f-963f-cbd235500b2e-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-572xx\" (UID: \"b0e181e1-d1fa-460f-963f-cbd235500b2e\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.570705 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx" Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.604520 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68"] Dec 04 04:07:31 crc kubenswrapper[4806]: I1204 04:07:31.793271 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7d545d4ccb-h5stl"] Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.084903 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.162888 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6df24743-9526-4457-9f48-f8639a96977f-utilities\") pod \"6df24743-9526-4457-9f48-f8639a96977f\" (UID: \"6df24743-9526-4457-9f48-f8639a96977f\") " Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.163042 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsbx6\" (UniqueName: \"kubernetes.io/projected/6df24743-9526-4457-9f48-f8639a96977f-kube-api-access-wsbx6\") pod \"6df24743-9526-4457-9f48-f8639a96977f\" (UID: \"6df24743-9526-4457-9f48-f8639a96977f\") " Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.163122 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6df24743-9526-4457-9f48-f8639a96977f-catalog-content\") pod \"6df24743-9526-4457-9f48-f8639a96977f\" (UID: \"6df24743-9526-4457-9f48-f8639a96977f\") " Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.164785 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6df24743-9526-4457-9f48-f8639a96977f-utilities" (OuterVolumeSpecName: "utilities") pod "6df24743-9526-4457-9f48-f8639a96977f" (UID: "6df24743-9526-4457-9f48-f8639a96977f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.181220 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6df24743-9526-4457-9f48-f8639a96977f-kube-api-access-wsbx6" (OuterVolumeSpecName: "kube-api-access-wsbx6") pod "6df24743-9526-4457-9f48-f8639a96977f" (UID: "6df24743-9526-4457-9f48-f8639a96977f"). InnerVolumeSpecName "kube-api-access-wsbx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.184713 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx"] Dec 04 04:07:32 crc kubenswrapper[4806]: W1204 04:07:32.197124 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0e181e1_d1fa_460f_963f_cbd235500b2e.slice/crio-dc8c48c321a381fc1a50e22e65471704d151b061b00a954c61e6a05a1fb48116 WatchSource:0}: Error finding container dc8c48c321a381fc1a50e22e65471704d151b061b00a954c61e6a05a1fb48116: Status 404 returned error can't find the container with id dc8c48c321a381fc1a50e22e65471704d151b061b00a954c61e6a05a1fb48116 Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.266752 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsbx6\" (UniqueName: \"kubernetes.io/projected/6df24743-9526-4457-9f48-f8639a96977f-kube-api-access-wsbx6\") on node \"crc\" DevicePath \"\"" Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.269363 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6df24743-9526-4457-9f48-f8639a96977f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.277315 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6df24743-9526-4457-9f48-f8639a96977f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6df24743-9526-4457-9f48-f8639a96977f" (UID: "6df24743-9526-4457-9f48-f8639a96977f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.312781 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68" event={"ID":"eae97ef2-3aeb-4303-b26f-0f84a75abda0","Type":"ContainerStarted","Data":"6f537e500d245e7e577f644af9c8770e12f5e3780ac0702e33a645f7bd21b569"} Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.314857 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7d545d4ccb-h5stl" event={"ID":"82cf8319-340a-4d1d-92e4-adfadea8a5cf","Type":"ContainerStarted","Data":"540902ceba624b38b182c5745f5a76851f2bcfb522bab9bf40300cb710fb01af"} Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.314946 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7d545d4ccb-h5stl" event={"ID":"82cf8319-340a-4d1d-92e4-adfadea8a5cf","Type":"ContainerStarted","Data":"dc5037ad5f1871bb6b2b66f7403be4b3b02f0cc160448ee705c49f50f19ae16e"} Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.316282 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx" event={"ID":"b0e181e1-d1fa-460f-963f-cbd235500b2e","Type":"ContainerStarted","Data":"dc8c48c321a381fc1a50e22e65471704d151b061b00a954c61e6a05a1fb48116"} Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.320950 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pwlns" event={"ID":"6df24743-9526-4457-9f48-f8639a96977f","Type":"ContainerDied","Data":"671ddf7e9e30c07a75c79ec10f657d6dde848606b7c9c384392c9ef1a43c5a48"} Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.321000 4806 scope.go:117] "RemoveContainer" containerID="f0ef149b28b6df5facad56f98fae2579aaf63d57cd97e0a096c849b1b054c1d5" Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.321049 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pwlns" Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.338959 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7d545d4ccb-h5stl" podStartSLOduration=2.338911305 podStartE2EDuration="2.338911305s" podCreationTimestamp="2025-12-04 04:07:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:07:32.336615174 +0000 UTC m=+767.195128122" watchObservedRunningTime="2025-12-04 04:07:32.338911305 +0000 UTC m=+767.197424233" Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.371085 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pwlns"] Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.373541 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6df24743-9526-4457-9f48-f8639a96977f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.373778 4806 scope.go:117] "RemoveContainer" containerID="8f6465e25ee5f249a999c4a24b9d50faf319818d217505569fc04aac34ff6f15" Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.376480 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pwlns"] Dec 04 04:07:32 crc kubenswrapper[4806]: I1204 04:07:32.405370 4806 scope.go:117] "RemoveContainer" containerID="fc96cfaa23666ab41da5aac3901e25a3328ec4e7febf98918543fd73ec85649c" Dec 04 04:07:33 crc kubenswrapper[4806]: I1204 04:07:33.433426 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6df24743-9526-4457-9f48-f8639a96977f" path="/var/lib/kubelet/pods/6df24743-9526-4457-9f48-f8639a96977f/volumes" Dec 04 04:07:34 crc kubenswrapper[4806]: I1204 04:07:34.357378 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-pdg7x" event={"ID":"843789a7-01bb-42c0-9f30-aa52af788147","Type":"ContainerStarted","Data":"b61b7bf2db2cc6ed93a7ac21494cf9e9a5f7d6fbab598a27d08f35e13e4629ee"} Dec 04 04:07:34 crc kubenswrapper[4806]: I1204 04:07:34.357686 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-pdg7x" Dec 04 04:07:34 crc kubenswrapper[4806]: I1204 04:07:34.362774 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68" event={"ID":"eae97ef2-3aeb-4303-b26f-0f84a75abda0","Type":"ContainerStarted","Data":"ccfd3c7c5f58c54b601664ed37080650fc3cb926b33bf4f302a66af2508d9032"} Dec 04 04:07:34 crc kubenswrapper[4806]: I1204 04:07:34.363820 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68" Dec 04 04:07:34 crc kubenswrapper[4806]: I1204 04:07:34.365434 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-dlkzr" event={"ID":"7b294b89-f409-40b7-8ba4-72e53da24a6f","Type":"ContainerStarted","Data":"25b0de8f87ebaf5483458726488129886d4c623d4ba8febdd470498b47fd9c9f"} Dec 04 04:07:34 crc kubenswrapper[4806]: I1204 04:07:34.381427 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-pdg7x" podStartSLOduration=1.424514085 podStartE2EDuration="4.381400987s" podCreationTimestamp="2025-12-04 04:07:30 +0000 UTC" firstStartedPulling="2025-12-04 04:07:30.823408254 +0000 UTC m=+765.681921182" lastFinishedPulling="2025-12-04 04:07:33.780295146 +0000 UTC m=+768.638808084" observedRunningTime="2025-12-04 04:07:34.379409506 +0000 UTC m=+769.237922434" watchObservedRunningTime="2025-12-04 04:07:34.381400987 +0000 UTC m=+769.239913915" Dec 04 04:07:34 crc kubenswrapper[4806]: I1204 04:07:34.399708 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68" podStartSLOduration=2.235938653 podStartE2EDuration="4.399674373s" podCreationTimestamp="2025-12-04 04:07:30 +0000 UTC" firstStartedPulling="2025-12-04 04:07:31.66013208 +0000 UTC m=+766.518645008" lastFinishedPulling="2025-12-04 04:07:33.8238678 +0000 UTC m=+768.682380728" observedRunningTime="2025-12-04 04:07:34.398531744 +0000 UTC m=+769.257044692" watchObservedRunningTime="2025-12-04 04:07:34.399674373 +0000 UTC m=+769.258187301" Dec 04 04:07:35 crc kubenswrapper[4806]: I1204 04:07:35.380698 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx" event={"ID":"b0e181e1-d1fa-460f-963f-cbd235500b2e","Type":"ContainerStarted","Data":"2b3f247a09cac70b900f99cace3cddf355b3e85f5acc95fe3b5ea7163d625fee"} Dec 04 04:07:35 crc kubenswrapper[4806]: I1204 04:07:35.404822 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-572xx" podStartSLOduration=2.581910562 podStartE2EDuration="5.404800825s" podCreationTimestamp="2025-12-04 04:07:30 +0000 UTC" firstStartedPulling="2025-12-04 04:07:32.201916227 +0000 UTC m=+767.060429155" lastFinishedPulling="2025-12-04 04:07:35.0248065 +0000 UTC m=+769.883319418" observedRunningTime="2025-12-04 04:07:35.404008195 +0000 UTC m=+770.262521133" watchObservedRunningTime="2025-12-04 04:07:35.404800825 +0000 UTC m=+770.263313753" Dec 04 04:07:37 crc kubenswrapper[4806]: I1204 04:07:37.398148 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-dlkzr" event={"ID":"7b294b89-f409-40b7-8ba4-72e53da24a6f","Type":"ContainerStarted","Data":"b05e60c0ed0ad3880ca807c63b57b09875a5a1ddb8d0305b3c4468b79ec728af"} Dec 04 04:07:37 crc kubenswrapper[4806]: I1204 04:07:37.423433 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-dlkzr" podStartSLOduration=2.322883457 podStartE2EDuration="7.423407496s" podCreationTimestamp="2025-12-04 04:07:30 +0000 UTC" firstStartedPulling="2025-12-04 04:07:31.151993169 +0000 UTC m=+766.010506107" lastFinishedPulling="2025-12-04 04:07:36.252517218 +0000 UTC m=+771.111030146" observedRunningTime="2025-12-04 04:07:37.416513867 +0000 UTC m=+772.275026795" watchObservedRunningTime="2025-12-04 04:07:37.423407496 +0000 UTC m=+772.281920424" Dec 04 04:07:40 crc kubenswrapper[4806]: I1204 04:07:40.792287 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-pdg7x" Dec 04 04:07:41 crc kubenswrapper[4806]: I1204 04:07:41.373457 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:41 crc kubenswrapper[4806]: I1204 04:07:41.373548 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:41 crc kubenswrapper[4806]: I1204 04:07:41.378558 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:41 crc kubenswrapper[4806]: I1204 04:07:41.432803 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7d545d4ccb-h5stl" Dec 04 04:07:41 crc kubenswrapper[4806]: I1204 04:07:41.498721 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-z2mdj"] Dec 04 04:07:51 crc kubenswrapper[4806]: I1204 04:07:51.350568 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-q9x68" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.550093 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4"] Dec 04 04:08:03 crc kubenswrapper[4806]: E1204 04:08:03.551780 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6df24743-9526-4457-9f48-f8639a96977f" containerName="registry-server" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.551849 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="6df24743-9526-4457-9f48-f8639a96977f" containerName="registry-server" Dec 04 04:08:03 crc kubenswrapper[4806]: E1204 04:08:03.551914 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6df24743-9526-4457-9f48-f8639a96977f" containerName="extract-utilities" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.552034 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="6df24743-9526-4457-9f48-f8639a96977f" containerName="extract-utilities" Dec 04 04:08:03 crc kubenswrapper[4806]: E1204 04:08:03.552098 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6df24743-9526-4457-9f48-f8639a96977f" containerName="extract-content" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.552150 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="6df24743-9526-4457-9f48-f8639a96977f" containerName="extract-content" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.552292 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="6df24743-9526-4457-9f48-f8639a96977f" containerName="registry-server" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.553040 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.556341 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.556784 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/205b1465-25a7-43a5-9449-78baf9a87116-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4\" (UID: \"205b1465-25a7-43a5-9449-78baf9a87116\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.556864 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8mn2\" (UniqueName: \"kubernetes.io/projected/205b1465-25a7-43a5-9449-78baf9a87116-kube-api-access-w8mn2\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4\" (UID: \"205b1465-25a7-43a5-9449-78baf9a87116\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.557005 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/205b1465-25a7-43a5-9449-78baf9a87116-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4\" (UID: \"205b1465-25a7-43a5-9449-78baf9a87116\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.574862 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4"] Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.658415 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/205b1465-25a7-43a5-9449-78baf9a87116-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4\" (UID: \"205b1465-25a7-43a5-9449-78baf9a87116\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.658592 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/205b1465-25a7-43a5-9449-78baf9a87116-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4\" (UID: \"205b1465-25a7-43a5-9449-78baf9a87116\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.658636 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8mn2\" (UniqueName: \"kubernetes.io/projected/205b1465-25a7-43a5-9449-78baf9a87116-kube-api-access-w8mn2\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4\" (UID: \"205b1465-25a7-43a5-9449-78baf9a87116\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.659150 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/205b1465-25a7-43a5-9449-78baf9a87116-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4\" (UID: \"205b1465-25a7-43a5-9449-78baf9a87116\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.659202 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/205b1465-25a7-43a5-9449-78baf9a87116-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4\" (UID: \"205b1465-25a7-43a5-9449-78baf9a87116\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.681872 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8mn2\" (UniqueName: \"kubernetes.io/projected/205b1465-25a7-43a5-9449-78baf9a87116-kube-api-access-w8mn2\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4\" (UID: \"205b1465-25a7-43a5-9449-78baf9a87116\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" Dec 04 04:08:03 crc kubenswrapper[4806]: I1204 04:08:03.869070 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" Dec 04 04:08:04 crc kubenswrapper[4806]: I1204 04:08:04.259196 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4"] Dec 04 04:08:04 crc kubenswrapper[4806]: I1204 04:08:04.578467 4806 generic.go:334] "Generic (PLEG): container finished" podID="205b1465-25a7-43a5-9449-78baf9a87116" containerID="e8444c71caee69485b2914ce976c5646d45c2b91bfc3adf1caffd85e1f7e2687" exitCode=0 Dec 04 04:08:04 crc kubenswrapper[4806]: I1204 04:08:04.578536 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" event={"ID":"205b1465-25a7-43a5-9449-78baf9a87116","Type":"ContainerDied","Data":"e8444c71caee69485b2914ce976c5646d45c2b91bfc3adf1caffd85e1f7e2687"} Dec 04 04:08:04 crc kubenswrapper[4806]: I1204 04:08:04.578581 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" event={"ID":"205b1465-25a7-43a5-9449-78baf9a87116","Type":"ContainerStarted","Data":"35c0c4bc187c56aca5ba5747235f7c4f257e3dcee1bfeafd558afd6611ab845c"} Dec 04 04:08:06 crc kubenswrapper[4806]: I1204 04:08:06.542083 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-z2mdj" podUID="a3f377d2-10db-41db-9361-06abff2e0d29" containerName="console" containerID="cri-o://283b3e846066c6ed47471e5a08b0db543f827e835d01a438071913c135915670" gracePeriod=15 Dec 04 04:08:06 crc kubenswrapper[4806]: I1204 04:08:06.594049 4806 generic.go:334] "Generic (PLEG): container finished" podID="205b1465-25a7-43a5-9449-78baf9a87116" containerID="bb30b9e67e1840b6a56f10beb5e21f4b3287c71594f85971cf60e6935c48ab70" exitCode=0 Dec 04 04:08:06 crc kubenswrapper[4806]: I1204 04:08:06.594109 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" event={"ID":"205b1465-25a7-43a5-9449-78baf9a87116","Type":"ContainerDied","Data":"bb30b9e67e1840b6a56f10beb5e21f4b3287c71594f85971cf60e6935c48ab70"} Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.014078 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-z2mdj_a3f377d2-10db-41db-9361-06abff2e0d29/console/0.log" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.014637 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.124415 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcvgf\" (UniqueName: \"kubernetes.io/projected/a3f377d2-10db-41db-9361-06abff2e0d29-kube-api-access-lcvgf\") pod \"a3f377d2-10db-41db-9361-06abff2e0d29\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.124501 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-service-ca\") pod \"a3f377d2-10db-41db-9361-06abff2e0d29\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.124586 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-console-config\") pod \"a3f377d2-10db-41db-9361-06abff2e0d29\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.124634 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a3f377d2-10db-41db-9361-06abff2e0d29-console-oauth-config\") pod \"a3f377d2-10db-41db-9361-06abff2e0d29\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.124660 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a3f377d2-10db-41db-9361-06abff2e0d29-console-serving-cert\") pod \"a3f377d2-10db-41db-9361-06abff2e0d29\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.124681 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-oauth-serving-cert\") pod \"a3f377d2-10db-41db-9361-06abff2e0d29\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.124701 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-trusted-ca-bundle\") pod \"a3f377d2-10db-41db-9361-06abff2e0d29\" (UID: \"a3f377d2-10db-41db-9361-06abff2e0d29\") " Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.125653 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "a3f377d2-10db-41db-9361-06abff2e0d29" (UID: "a3f377d2-10db-41db-9361-06abff2e0d29"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.125994 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "a3f377d2-10db-41db-9361-06abff2e0d29" (UID: "a3f377d2-10db-41db-9361-06abff2e0d29"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.126011 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-console-config" (OuterVolumeSpecName: "console-config") pod "a3f377d2-10db-41db-9361-06abff2e0d29" (UID: "a3f377d2-10db-41db-9361-06abff2e0d29"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.126657 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-service-ca" (OuterVolumeSpecName: "service-ca") pod "a3f377d2-10db-41db-9361-06abff2e0d29" (UID: "a3f377d2-10db-41db-9361-06abff2e0d29"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.131879 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3f377d2-10db-41db-9361-06abff2e0d29-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "a3f377d2-10db-41db-9361-06abff2e0d29" (UID: "a3f377d2-10db-41db-9361-06abff2e0d29"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.131971 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3f377d2-10db-41db-9361-06abff2e0d29-kube-api-access-lcvgf" (OuterVolumeSpecName: "kube-api-access-lcvgf") pod "a3f377d2-10db-41db-9361-06abff2e0d29" (UID: "a3f377d2-10db-41db-9361-06abff2e0d29"). InnerVolumeSpecName "kube-api-access-lcvgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.134544 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3f377d2-10db-41db-9361-06abff2e0d29-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "a3f377d2-10db-41db-9361-06abff2e0d29" (UID: "a3f377d2-10db-41db-9361-06abff2e0d29"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.226164 4806 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-console-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.226209 4806 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a3f377d2-10db-41db-9361-06abff2e0d29-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.226226 4806 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a3f377d2-10db-41db-9361-06abff2e0d29-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.226235 4806 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.226245 4806 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.226254 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcvgf\" (UniqueName: \"kubernetes.io/projected/a3f377d2-10db-41db-9361-06abff2e0d29-kube-api-access-lcvgf\") on node \"crc\" DevicePath \"\"" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.226266 4806 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a3f377d2-10db-41db-9361-06abff2e0d29-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.604630 4806 generic.go:334] "Generic (PLEG): container finished" podID="205b1465-25a7-43a5-9449-78baf9a87116" containerID="a867b3372b7f907d1bdad1a540ea50cd0e0fdb92a3a8548fa89a18258c73c235" exitCode=0 Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.604795 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" event={"ID":"205b1465-25a7-43a5-9449-78baf9a87116","Type":"ContainerDied","Data":"a867b3372b7f907d1bdad1a540ea50cd0e0fdb92a3a8548fa89a18258c73c235"} Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.609220 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-z2mdj_a3f377d2-10db-41db-9361-06abff2e0d29/console/0.log" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.609335 4806 generic.go:334] "Generic (PLEG): container finished" podID="a3f377d2-10db-41db-9361-06abff2e0d29" containerID="283b3e846066c6ed47471e5a08b0db543f827e835d01a438071913c135915670" exitCode=2 Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.609449 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-z2mdj" event={"ID":"a3f377d2-10db-41db-9361-06abff2e0d29","Type":"ContainerDied","Data":"283b3e846066c6ed47471e5a08b0db543f827e835d01a438071913c135915670"} Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.609532 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-z2mdj" event={"ID":"a3f377d2-10db-41db-9361-06abff2e0d29","Type":"ContainerDied","Data":"ebf71e78512c42d1383b5fd5e7784665ac98983fced4976925e418cadb212c4e"} Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.609606 4806 scope.go:117] "RemoveContainer" containerID="283b3e846066c6ed47471e5a08b0db543f827e835d01a438071913c135915670" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.609809 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-z2mdj" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.640908 4806 scope.go:117] "RemoveContainer" containerID="283b3e846066c6ed47471e5a08b0db543f827e835d01a438071913c135915670" Dec 04 04:08:07 crc kubenswrapper[4806]: E1204 04:08:07.642447 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"283b3e846066c6ed47471e5a08b0db543f827e835d01a438071913c135915670\": container with ID starting with 283b3e846066c6ed47471e5a08b0db543f827e835d01a438071913c135915670 not found: ID does not exist" containerID="283b3e846066c6ed47471e5a08b0db543f827e835d01a438071913c135915670" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.642486 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"283b3e846066c6ed47471e5a08b0db543f827e835d01a438071913c135915670"} err="failed to get container status \"283b3e846066c6ed47471e5a08b0db543f827e835d01a438071913c135915670\": rpc error: code = NotFound desc = could not find container \"283b3e846066c6ed47471e5a08b0db543f827e835d01a438071913c135915670\": container with ID starting with 283b3e846066c6ed47471e5a08b0db543f827e835d01a438071913c135915670 not found: ID does not exist" Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.648348 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-z2mdj"] Dec 04 04:08:07 crc kubenswrapper[4806]: I1204 04:08:07.652336 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-z2mdj"] Dec 04 04:08:08 crc kubenswrapper[4806]: I1204 04:08:08.859500 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" Dec 04 04:08:08 crc kubenswrapper[4806]: I1204 04:08:08.985098 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/205b1465-25a7-43a5-9449-78baf9a87116-bundle\") pod \"205b1465-25a7-43a5-9449-78baf9a87116\" (UID: \"205b1465-25a7-43a5-9449-78baf9a87116\") " Dec 04 04:08:08 crc kubenswrapper[4806]: I1204 04:08:08.985228 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/205b1465-25a7-43a5-9449-78baf9a87116-util\") pod \"205b1465-25a7-43a5-9449-78baf9a87116\" (UID: \"205b1465-25a7-43a5-9449-78baf9a87116\") " Dec 04 04:08:08 crc kubenswrapper[4806]: I1204 04:08:08.985277 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8mn2\" (UniqueName: \"kubernetes.io/projected/205b1465-25a7-43a5-9449-78baf9a87116-kube-api-access-w8mn2\") pod \"205b1465-25a7-43a5-9449-78baf9a87116\" (UID: \"205b1465-25a7-43a5-9449-78baf9a87116\") " Dec 04 04:08:08 crc kubenswrapper[4806]: I1204 04:08:08.987502 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/205b1465-25a7-43a5-9449-78baf9a87116-bundle" (OuterVolumeSpecName: "bundle") pod "205b1465-25a7-43a5-9449-78baf9a87116" (UID: "205b1465-25a7-43a5-9449-78baf9a87116"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:08:08 crc kubenswrapper[4806]: I1204 04:08:08.994554 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/205b1465-25a7-43a5-9449-78baf9a87116-kube-api-access-w8mn2" (OuterVolumeSpecName: "kube-api-access-w8mn2") pod "205b1465-25a7-43a5-9449-78baf9a87116" (UID: "205b1465-25a7-43a5-9449-78baf9a87116"). InnerVolumeSpecName "kube-api-access-w8mn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:08:09 crc kubenswrapper[4806]: I1204 04:08:09.023616 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/205b1465-25a7-43a5-9449-78baf9a87116-util" (OuterVolumeSpecName: "util") pod "205b1465-25a7-43a5-9449-78baf9a87116" (UID: "205b1465-25a7-43a5-9449-78baf9a87116"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:08:09 crc kubenswrapper[4806]: I1204 04:08:09.086794 4806 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/205b1465-25a7-43a5-9449-78baf9a87116-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:08:09 crc kubenswrapper[4806]: I1204 04:08:09.086838 4806 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/205b1465-25a7-43a5-9449-78baf9a87116-util\") on node \"crc\" DevicePath \"\"" Dec 04 04:08:09 crc kubenswrapper[4806]: I1204 04:08:09.086849 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8mn2\" (UniqueName: \"kubernetes.io/projected/205b1465-25a7-43a5-9449-78baf9a87116-kube-api-access-w8mn2\") on node \"crc\" DevicePath \"\"" Dec 04 04:08:09 crc kubenswrapper[4806]: I1204 04:08:09.432219 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3f377d2-10db-41db-9361-06abff2e0d29" path="/var/lib/kubelet/pods/a3f377d2-10db-41db-9361-06abff2e0d29/volumes" Dec 04 04:08:09 crc kubenswrapper[4806]: I1204 04:08:09.635722 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" event={"ID":"205b1465-25a7-43a5-9449-78baf9a87116","Type":"ContainerDied","Data":"35c0c4bc187c56aca5ba5747235f7c4f257e3dcee1bfeafd558afd6611ab845c"} Dec 04 04:08:09 crc kubenswrapper[4806]: I1204 04:08:09.636278 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35c0c4bc187c56aca5ba5747235f7c4f257e3dcee1bfeafd558afd6611ab845c" Dec 04 04:08:09 crc kubenswrapper[4806]: I1204 04:08:09.635838 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.191986 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh"] Dec 04 04:08:17 crc kubenswrapper[4806]: E1204 04:08:17.193202 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="205b1465-25a7-43a5-9449-78baf9a87116" containerName="pull" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.193222 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="205b1465-25a7-43a5-9449-78baf9a87116" containerName="pull" Dec 04 04:08:17 crc kubenswrapper[4806]: E1204 04:08:17.193244 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3f377d2-10db-41db-9361-06abff2e0d29" containerName="console" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.193250 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3f377d2-10db-41db-9361-06abff2e0d29" containerName="console" Dec 04 04:08:17 crc kubenswrapper[4806]: E1204 04:08:17.193268 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="205b1465-25a7-43a5-9449-78baf9a87116" containerName="util" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.193278 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="205b1465-25a7-43a5-9449-78baf9a87116" containerName="util" Dec 04 04:08:17 crc kubenswrapper[4806]: E1204 04:08:17.193288 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="205b1465-25a7-43a5-9449-78baf9a87116" containerName="extract" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.193294 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="205b1465-25a7-43a5-9449-78baf9a87116" containerName="extract" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.193399 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3f377d2-10db-41db-9361-06abff2e0d29" containerName="console" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.193414 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="205b1465-25a7-43a5-9449-78baf9a87116" containerName="extract" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.194021 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" Dec 04 04:08:17 crc kubenswrapper[4806]: W1204 04:08:17.201036 4806 reflector.go:561] object-"metallb-system"/"metallb-operator-webhook-server-cert": failed to list *v1.Secret: secrets "metallb-operator-webhook-server-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Dec 04 04:08:17 crc kubenswrapper[4806]: W1204 04:08:17.201070 4806 reflector.go:561] object-"metallb-system"/"metallb-operator-controller-manager-service-cert": failed to list *v1.Secret: secrets "metallb-operator-controller-manager-service-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Dec 04 04:08:17 crc kubenswrapper[4806]: E1204 04:08:17.201100 4806 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-operator-controller-manager-service-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-operator-controller-manager-service-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 04:08:17 crc kubenswrapper[4806]: E1204 04:08:17.201102 4806 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-operator-webhook-server-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-operator-webhook-server-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 04:08:17 crc kubenswrapper[4806]: W1204 04:08:17.201036 4806 reflector.go:561] object-"metallb-system"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Dec 04 04:08:17 crc kubenswrapper[4806]: E1204 04:08:17.201133 4806 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 04:08:17 crc kubenswrapper[4806]: W1204 04:08:17.201148 4806 reflector.go:561] object-"metallb-system"/"manager-account-dockercfg-d24tv": failed to list *v1.Secret: secrets "manager-account-dockercfg-d24tv" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Dec 04 04:08:17 crc kubenswrapper[4806]: E1204 04:08:17.201227 4806 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"manager-account-dockercfg-d24tv\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"manager-account-dockercfg-d24tv\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 04:08:17 crc kubenswrapper[4806]: W1204 04:08:17.201639 4806 reflector.go:561] object-"metallb-system"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Dec 04 04:08:17 crc kubenswrapper[4806]: E1204 04:08:17.201809 4806 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.268356 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh"] Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.271507 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/09d57bc9-a5d5-4848-9cbd-12ce41ca2b78-apiservice-cert\") pod \"metallb-operator-controller-manager-5b99fcbdc-n6szh\" (UID: \"09d57bc9-a5d5-4848-9cbd-12ce41ca2b78\") " pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.271573 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqzcv\" (UniqueName: \"kubernetes.io/projected/09d57bc9-a5d5-4848-9cbd-12ce41ca2b78-kube-api-access-cqzcv\") pod \"metallb-operator-controller-manager-5b99fcbdc-n6szh\" (UID: \"09d57bc9-a5d5-4848-9cbd-12ce41ca2b78\") " pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.271627 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/09d57bc9-a5d5-4848-9cbd-12ce41ca2b78-webhook-cert\") pod \"metallb-operator-controller-manager-5b99fcbdc-n6szh\" (UID: \"09d57bc9-a5d5-4848-9cbd-12ce41ca2b78\") " pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.373746 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/09d57bc9-a5d5-4848-9cbd-12ce41ca2b78-apiservice-cert\") pod \"metallb-operator-controller-manager-5b99fcbdc-n6szh\" (UID: \"09d57bc9-a5d5-4848-9cbd-12ce41ca2b78\") " pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.374347 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqzcv\" (UniqueName: \"kubernetes.io/projected/09d57bc9-a5d5-4848-9cbd-12ce41ca2b78-kube-api-access-cqzcv\") pod \"metallb-operator-controller-manager-5b99fcbdc-n6szh\" (UID: \"09d57bc9-a5d5-4848-9cbd-12ce41ca2b78\") " pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.374418 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/09d57bc9-a5d5-4848-9cbd-12ce41ca2b78-webhook-cert\") pod \"metallb-operator-controller-manager-5b99fcbdc-n6szh\" (UID: \"09d57bc9-a5d5-4848-9cbd-12ce41ca2b78\") " pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.719328 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp"] Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.720939 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.723290 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.723960 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-hmnqb" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.724974 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.780660 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/134cc417-ef0e-4cb7-90e5-5a8f55466f6a-apiservice-cert\") pod \"metallb-operator-webhook-server-58dfcfbb98-7prrp\" (UID: \"134cc417-ef0e-4cb7-90e5-5a8f55466f6a\") " pod="metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.780726 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/134cc417-ef0e-4cb7-90e5-5a8f55466f6a-webhook-cert\") pod \"metallb-operator-webhook-server-58dfcfbb98-7prrp\" (UID: \"134cc417-ef0e-4cb7-90e5-5a8f55466f6a\") " pod="metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.780815 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxxlw\" (UniqueName: \"kubernetes.io/projected/134cc417-ef0e-4cb7-90e5-5a8f55466f6a-kube-api-access-mxxlw\") pod \"metallb-operator-webhook-server-58dfcfbb98-7prrp\" (UID: \"134cc417-ef0e-4cb7-90e5-5a8f55466f6a\") " pod="metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.802521 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp"] Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.883195 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/134cc417-ef0e-4cb7-90e5-5a8f55466f6a-apiservice-cert\") pod \"metallb-operator-webhook-server-58dfcfbb98-7prrp\" (UID: \"134cc417-ef0e-4cb7-90e5-5a8f55466f6a\") " pod="metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.883267 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/134cc417-ef0e-4cb7-90e5-5a8f55466f6a-webhook-cert\") pod \"metallb-operator-webhook-server-58dfcfbb98-7prrp\" (UID: \"134cc417-ef0e-4cb7-90e5-5a8f55466f6a\") " pod="metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.883590 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxxlw\" (UniqueName: \"kubernetes.io/projected/134cc417-ef0e-4cb7-90e5-5a8f55466f6a-kube-api-access-mxxlw\") pod \"metallb-operator-webhook-server-58dfcfbb98-7prrp\" (UID: \"134cc417-ef0e-4cb7-90e5-5a8f55466f6a\") " pod="metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.890741 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/134cc417-ef0e-4cb7-90e5-5a8f55466f6a-apiservice-cert\") pod \"metallb-operator-webhook-server-58dfcfbb98-7prrp\" (UID: \"134cc417-ef0e-4cb7-90e5-5a8f55466f6a\") " pod="metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp" Dec 04 04:08:17 crc kubenswrapper[4806]: I1204 04:08:17.890754 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/134cc417-ef0e-4cb7-90e5-5a8f55466f6a-webhook-cert\") pod \"metallb-operator-webhook-server-58dfcfbb98-7prrp\" (UID: \"134cc417-ef0e-4cb7-90e5-5a8f55466f6a\") " pod="metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp" Dec 04 04:08:18 crc kubenswrapper[4806]: I1204 04:08:18.007436 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-d24tv" Dec 04 04:08:18 crc kubenswrapper[4806]: E1204 04:08:18.374823 4806 secret.go:188] Couldn't get secret metallb-system/metallb-operator-controller-manager-service-cert: failed to sync secret cache: timed out waiting for the condition Dec 04 04:08:18 crc kubenswrapper[4806]: E1204 04:08:18.374823 4806 secret.go:188] Couldn't get secret metallb-system/metallb-operator-controller-manager-service-cert: failed to sync secret cache: timed out waiting for the condition Dec 04 04:08:18 crc kubenswrapper[4806]: E1204 04:08:18.374952 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09d57bc9-a5d5-4848-9cbd-12ce41ca2b78-apiservice-cert podName:09d57bc9-a5d5-4848-9cbd-12ce41ca2b78 nodeName:}" failed. No retries permitted until 2025-12-04 04:08:18.87491231 +0000 UTC m=+813.733425238 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/09d57bc9-a5d5-4848-9cbd-12ce41ca2b78-apiservice-cert") pod "metallb-operator-controller-manager-5b99fcbdc-n6szh" (UID: "09d57bc9-a5d5-4848-9cbd-12ce41ca2b78") : failed to sync secret cache: timed out waiting for the condition Dec 04 04:08:18 crc kubenswrapper[4806]: E1204 04:08:18.375052 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/09d57bc9-a5d5-4848-9cbd-12ce41ca2b78-webhook-cert podName:09d57bc9-a5d5-4848-9cbd-12ce41ca2b78 nodeName:}" failed. No retries permitted until 2025-12-04 04:08:18.874998833 +0000 UTC m=+813.733511761 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/09d57bc9-a5d5-4848-9cbd-12ce41ca2b78-webhook-cert") pod "metallb-operator-controller-manager-5b99fcbdc-n6szh" (UID: "09d57bc9-a5d5-4848-9cbd-12ce41ca2b78") : failed to sync secret cache: timed out waiting for the condition Dec 04 04:08:18 crc kubenswrapper[4806]: I1204 04:08:18.375158 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 04 04:08:18 crc kubenswrapper[4806]: I1204 04:08:18.581062 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 04 04:08:18 crc kubenswrapper[4806]: I1204 04:08:18.593688 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqzcv\" (UniqueName: \"kubernetes.io/projected/09d57bc9-a5d5-4848-9cbd-12ce41ca2b78-kube-api-access-cqzcv\") pod \"metallb-operator-controller-manager-5b99fcbdc-n6szh\" (UID: \"09d57bc9-a5d5-4848-9cbd-12ce41ca2b78\") " pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" Dec 04 04:08:18 crc kubenswrapper[4806]: I1204 04:08:18.598027 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 04 04:08:18 crc kubenswrapper[4806]: I1204 04:08:18.613779 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxxlw\" (UniqueName: \"kubernetes.io/projected/134cc417-ef0e-4cb7-90e5-5a8f55466f6a-kube-api-access-mxxlw\") pod \"metallb-operator-webhook-server-58dfcfbb98-7prrp\" (UID: \"134cc417-ef0e-4cb7-90e5-5a8f55466f6a\") " pod="metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp" Dec 04 04:08:18 crc kubenswrapper[4806]: I1204 04:08:18.638365 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp" Dec 04 04:08:18 crc kubenswrapper[4806]: I1204 04:08:18.818975 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 04 04:08:18 crc kubenswrapper[4806]: I1204 04:08:18.917959 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/09d57bc9-a5d5-4848-9cbd-12ce41ca2b78-apiservice-cert\") pod \"metallb-operator-controller-manager-5b99fcbdc-n6szh\" (UID: \"09d57bc9-a5d5-4848-9cbd-12ce41ca2b78\") " pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" Dec 04 04:08:18 crc kubenswrapper[4806]: I1204 04:08:18.918065 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/09d57bc9-a5d5-4848-9cbd-12ce41ca2b78-webhook-cert\") pod \"metallb-operator-controller-manager-5b99fcbdc-n6szh\" (UID: \"09d57bc9-a5d5-4848-9cbd-12ce41ca2b78\") " pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" Dec 04 04:08:18 crc kubenswrapper[4806]: I1204 04:08:18.923290 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/09d57bc9-a5d5-4848-9cbd-12ce41ca2b78-apiservice-cert\") pod \"metallb-operator-controller-manager-5b99fcbdc-n6szh\" (UID: \"09d57bc9-a5d5-4848-9cbd-12ce41ca2b78\") " pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" Dec 04 04:08:18 crc kubenswrapper[4806]: I1204 04:08:18.923474 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/09d57bc9-a5d5-4848-9cbd-12ce41ca2b78-webhook-cert\") pod \"metallb-operator-controller-manager-5b99fcbdc-n6szh\" (UID: \"09d57bc9-a5d5-4848-9cbd-12ce41ca2b78\") " pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" Dec 04 04:08:19 crc kubenswrapper[4806]: I1204 04:08:19.014660 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" Dec 04 04:08:20 crc kubenswrapper[4806]: I1204 04:08:20.070003 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp"] Dec 04 04:08:20 crc kubenswrapper[4806]: I1204 04:08:20.141289 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh"] Dec 04 04:08:20 crc kubenswrapper[4806]: W1204 04:08:20.141965 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09d57bc9_a5d5_4848_9cbd_12ce41ca2b78.slice/crio-e9fa7bb1f00a7688c27b93e50058932148378fb29ef0e0c1ff7280eaf8364601 WatchSource:0}: Error finding container e9fa7bb1f00a7688c27b93e50058932148378fb29ef0e0c1ff7280eaf8364601: Status 404 returned error can't find the container with id e9fa7bb1f00a7688c27b93e50058932148378fb29ef0e0c1ff7280eaf8364601 Dec 04 04:08:20 crc kubenswrapper[4806]: I1204 04:08:20.837214 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp" event={"ID":"134cc417-ef0e-4cb7-90e5-5a8f55466f6a","Type":"ContainerStarted","Data":"65a837f97a75f7bea8949906b46dbefcf51c402196c33ce7f63be5f52f4ec21d"} Dec 04 04:08:20 crc kubenswrapper[4806]: I1204 04:08:20.845296 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" event={"ID":"09d57bc9-a5d5-4848-9cbd-12ce41ca2b78","Type":"ContainerStarted","Data":"e9fa7bb1f00a7688c27b93e50058932148378fb29ef0e0c1ff7280eaf8364601"} Dec 04 04:08:32 crc kubenswrapper[4806]: I1204 04:08:32.061834 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" event={"ID":"09d57bc9-a5d5-4848-9cbd-12ce41ca2b78","Type":"ContainerStarted","Data":"7852e4b793492a0e0b3986234ba625f1f4fa5ae471aacaa9d949d7c8e87bbed9"} Dec 04 04:08:32 crc kubenswrapper[4806]: I1204 04:08:32.062792 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" Dec 04 04:08:32 crc kubenswrapper[4806]: I1204 04:08:32.064951 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp" event={"ID":"134cc417-ef0e-4cb7-90e5-5a8f55466f6a","Type":"ContainerStarted","Data":"0e403f08b33ccc1c79dbed9bb82c422afb2cb94a435599ddbc68c32ecd83565a"} Dec 04 04:08:32 crc kubenswrapper[4806]: I1204 04:08:32.065109 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp" Dec 04 04:08:32 crc kubenswrapper[4806]: I1204 04:08:32.115789 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp" podStartSLOduration=4.008033803 podStartE2EDuration="15.115765506s" podCreationTimestamp="2025-12-04 04:08:17 +0000 UTC" firstStartedPulling="2025-12-04 04:08:20.088810613 +0000 UTC m=+814.947323541" lastFinishedPulling="2025-12-04 04:08:31.196542316 +0000 UTC m=+826.055055244" observedRunningTime="2025-12-04 04:08:32.115247351 +0000 UTC m=+826.973760289" watchObservedRunningTime="2025-12-04 04:08:32.115765506 +0000 UTC m=+826.974278434" Dec 04 04:08:32 crc kubenswrapper[4806]: I1204 04:08:32.118355 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" podStartSLOduration=4.087869672 podStartE2EDuration="15.118348577s" podCreationTimestamp="2025-12-04 04:08:17 +0000 UTC" firstStartedPulling="2025-12-04 04:08:20.145636102 +0000 UTC m=+815.004149020" lastFinishedPulling="2025-12-04 04:08:31.176114997 +0000 UTC m=+826.034627925" observedRunningTime="2025-12-04 04:08:32.096411981 +0000 UTC m=+826.954924899" watchObservedRunningTime="2025-12-04 04:08:32.118348577 +0000 UTC m=+826.976861505" Dec 04 04:08:48 crc kubenswrapper[4806]: I1204 04:08:48.646279 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-58dfcfbb98-7prrp" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.019758 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5b99fcbdc-n6szh" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.731337 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-6r5wz"] Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.734968 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.741577 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-9n6cm" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.741866 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.751544 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.771072 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/cb278465-e6d8-48b2-bc17-861968bba45e-frr-sockets\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.771133 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/cb278465-e6d8-48b2-bc17-861968bba45e-reloader\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.771156 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf4w2\" (UniqueName: \"kubernetes.io/projected/cb278465-e6d8-48b2-bc17-861968bba45e-kube-api-access-qf4w2\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.771198 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/cb278465-e6d8-48b2-bc17-861968bba45e-frr-conf\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.771228 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cb278465-e6d8-48b2-bc17-861968bba45e-metrics-certs\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.771250 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/cb278465-e6d8-48b2-bc17-861968bba45e-frr-startup\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.771267 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/cb278465-e6d8-48b2-bc17-861968bba45e-metrics\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.773702 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624"] Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.774967 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.782394 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.808867 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624"] Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.872608 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/cb278465-e6d8-48b2-bc17-861968bba45e-reloader\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.872696 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4rdd\" (UniqueName: \"kubernetes.io/projected/fac17a35-6b82-4397-855c-a9d2b27bcca6-kube-api-access-p4rdd\") pod \"frr-k8s-webhook-server-7fcb986d4-wh624\" (UID: \"fac17a35-6b82-4397-855c-a9d2b27bcca6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.872749 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf4w2\" (UniqueName: \"kubernetes.io/projected/cb278465-e6d8-48b2-bc17-861968bba45e-kube-api-access-qf4w2\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.872870 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fac17a35-6b82-4397-855c-a9d2b27bcca6-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-wh624\" (UID: \"fac17a35-6b82-4397-855c-a9d2b27bcca6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.872941 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/cb278465-e6d8-48b2-bc17-861968bba45e-frr-conf\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.872986 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cb278465-e6d8-48b2-bc17-861968bba45e-metrics-certs\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.873014 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/cb278465-e6d8-48b2-bc17-861968bba45e-frr-startup\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.873034 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/cb278465-e6d8-48b2-bc17-861968bba45e-metrics\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.873063 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/cb278465-e6d8-48b2-bc17-861968bba45e-frr-sockets\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.873189 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/cb278465-e6d8-48b2-bc17-861968bba45e-reloader\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.873382 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/cb278465-e6d8-48b2-bc17-861968bba45e-frr-sockets\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.873641 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/cb278465-e6d8-48b2-bc17-861968bba45e-frr-conf\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.873961 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/cb278465-e6d8-48b2-bc17-861968bba45e-metrics\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.874296 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/cb278465-e6d8-48b2-bc17-861968bba45e-frr-startup\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.893119 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cb278465-e6d8-48b2-bc17-861968bba45e-metrics-certs\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.897074 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf4w2\" (UniqueName: \"kubernetes.io/projected/cb278465-e6d8-48b2-bc17-861968bba45e-kube-api-access-qf4w2\") pod \"frr-k8s-6r5wz\" (UID: \"cb278465-e6d8-48b2-bc17-861968bba45e\") " pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.911299 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-g44sf"] Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.912425 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-g44sf" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.915404 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.915715 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.915874 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-97s96" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.917944 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.954240 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-qq2x8"] Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.955658 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-qq2x8" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.958508 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.974430 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-metallb-excludel2\") pod \"speaker-g44sf\" (UID: \"e653dfc5-42b0-4ae3-a622-b64b556d3ccc\") " pod="metallb-system/speaker-g44sf" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.974500 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkcpf\" (UniqueName: \"kubernetes.io/projected/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-kube-api-access-zkcpf\") pod \"speaker-g44sf\" (UID: \"e653dfc5-42b0-4ae3-a622-b64b556d3ccc\") " pod="metallb-system/speaker-g44sf" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.974583 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-memberlist\") pod \"speaker-g44sf\" (UID: \"e653dfc5-42b0-4ae3-a622-b64b556d3ccc\") " pod="metallb-system/speaker-g44sf" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.974605 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d96h2\" (UniqueName: \"kubernetes.io/projected/d80ba885-bdec-472c-b9eb-23928eeb6ab1-kube-api-access-d96h2\") pod \"controller-f8648f98b-qq2x8\" (UID: \"d80ba885-bdec-472c-b9eb-23928eeb6ab1\") " pod="metallb-system/controller-f8648f98b-qq2x8" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.974634 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d80ba885-bdec-472c-b9eb-23928eeb6ab1-cert\") pod \"controller-f8648f98b-qq2x8\" (UID: \"d80ba885-bdec-472c-b9eb-23928eeb6ab1\") " pod="metallb-system/controller-f8648f98b-qq2x8" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.974663 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4rdd\" (UniqueName: \"kubernetes.io/projected/fac17a35-6b82-4397-855c-a9d2b27bcca6-kube-api-access-p4rdd\") pod \"frr-k8s-webhook-server-7fcb986d4-wh624\" (UID: \"fac17a35-6b82-4397-855c-a9d2b27bcca6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.974965 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d80ba885-bdec-472c-b9eb-23928eeb6ab1-metrics-certs\") pod \"controller-f8648f98b-qq2x8\" (UID: \"d80ba885-bdec-472c-b9eb-23928eeb6ab1\") " pod="metallb-system/controller-f8648f98b-qq2x8" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.975038 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-metrics-certs\") pod \"speaker-g44sf\" (UID: \"e653dfc5-42b0-4ae3-a622-b64b556d3ccc\") " pod="metallb-system/speaker-g44sf" Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.975210 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fac17a35-6b82-4397-855c-a9d2b27bcca6-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-wh624\" (UID: \"fac17a35-6b82-4397-855c-a9d2b27bcca6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624" Dec 04 04:09:09 crc kubenswrapper[4806]: E1204 04:09:09.975359 4806 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 04 04:09:09 crc kubenswrapper[4806]: E1204 04:09:09.975437 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fac17a35-6b82-4397-855c-a9d2b27bcca6-cert podName:fac17a35-6b82-4397-855c-a9d2b27bcca6 nodeName:}" failed. No retries permitted until 2025-12-04 04:09:10.475408439 +0000 UTC m=+865.333921367 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fac17a35-6b82-4397-855c-a9d2b27bcca6-cert") pod "frr-k8s-webhook-server-7fcb986d4-wh624" (UID: "fac17a35-6b82-4397-855c-a9d2b27bcca6") : secret "frr-k8s-webhook-server-cert" not found Dec 04 04:09:09 crc kubenswrapper[4806]: I1204 04:09:09.981579 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-qq2x8"] Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:09.995999 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4rdd\" (UniqueName: \"kubernetes.io/projected/fac17a35-6b82-4397-855c-a9d2b27bcca6-kube-api-access-p4rdd\") pod \"frr-k8s-webhook-server-7fcb986d4-wh624\" (UID: \"fac17a35-6b82-4397-855c-a9d2b27bcca6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.057520 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.076070 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-metallb-excludel2\") pod \"speaker-g44sf\" (UID: \"e653dfc5-42b0-4ae3-a622-b64b556d3ccc\") " pod="metallb-system/speaker-g44sf" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.076130 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkcpf\" (UniqueName: \"kubernetes.io/projected/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-kube-api-access-zkcpf\") pod \"speaker-g44sf\" (UID: \"e653dfc5-42b0-4ae3-a622-b64b556d3ccc\") " pod="metallb-system/speaker-g44sf" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.076163 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-memberlist\") pod \"speaker-g44sf\" (UID: \"e653dfc5-42b0-4ae3-a622-b64b556d3ccc\") " pod="metallb-system/speaker-g44sf" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.076183 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d96h2\" (UniqueName: \"kubernetes.io/projected/d80ba885-bdec-472c-b9eb-23928eeb6ab1-kube-api-access-d96h2\") pod \"controller-f8648f98b-qq2x8\" (UID: \"d80ba885-bdec-472c-b9eb-23928eeb6ab1\") " pod="metallb-system/controller-f8648f98b-qq2x8" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.076210 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d80ba885-bdec-472c-b9eb-23928eeb6ab1-cert\") pod \"controller-f8648f98b-qq2x8\" (UID: \"d80ba885-bdec-472c-b9eb-23928eeb6ab1\") " pod="metallb-system/controller-f8648f98b-qq2x8" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.076239 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d80ba885-bdec-472c-b9eb-23928eeb6ab1-metrics-certs\") pod \"controller-f8648f98b-qq2x8\" (UID: \"d80ba885-bdec-472c-b9eb-23928eeb6ab1\") " pod="metallb-system/controller-f8648f98b-qq2x8" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.076270 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-metrics-certs\") pod \"speaker-g44sf\" (UID: \"e653dfc5-42b0-4ae3-a622-b64b556d3ccc\") " pod="metallb-system/speaker-g44sf" Dec 04 04:09:10 crc kubenswrapper[4806]: E1204 04:09:10.076383 4806 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 04 04:09:10 crc kubenswrapper[4806]: E1204 04:09:10.076443 4806 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 04 04:09:10 crc kubenswrapper[4806]: E1204 04:09:10.076451 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-metrics-certs podName:e653dfc5-42b0-4ae3-a622-b64b556d3ccc nodeName:}" failed. No retries permitted until 2025-12-04 04:09:10.576430432 +0000 UTC m=+865.434943360 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-metrics-certs") pod "speaker-g44sf" (UID: "e653dfc5-42b0-4ae3-a622-b64b556d3ccc") : secret "speaker-certs-secret" not found Dec 04 04:09:10 crc kubenswrapper[4806]: E1204 04:09:10.076590 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-memberlist podName:e653dfc5-42b0-4ae3-a622-b64b556d3ccc nodeName:}" failed. No retries permitted until 2025-12-04 04:09:10.576562856 +0000 UTC m=+865.435075784 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-memberlist") pod "speaker-g44sf" (UID: "e653dfc5-42b0-4ae3-a622-b64b556d3ccc") : secret "metallb-memberlist" not found Dec 04 04:09:10 crc kubenswrapper[4806]: E1204 04:09:10.076727 4806 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 04 04:09:10 crc kubenswrapper[4806]: E1204 04:09:10.076822 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d80ba885-bdec-472c-b9eb-23928eeb6ab1-metrics-certs podName:d80ba885-bdec-472c-b9eb-23928eeb6ab1 nodeName:}" failed. No retries permitted until 2025-12-04 04:09:10.576795584 +0000 UTC m=+865.435308512 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d80ba885-bdec-472c-b9eb-23928eeb6ab1-metrics-certs") pod "controller-f8648f98b-qq2x8" (UID: "d80ba885-bdec-472c-b9eb-23928eeb6ab1") : secret "controller-certs-secret" not found Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.077300 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-metallb-excludel2\") pod \"speaker-g44sf\" (UID: \"e653dfc5-42b0-4ae3-a622-b64b556d3ccc\") " pod="metallb-system/speaker-g44sf" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.086044 4806 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.103631 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d96h2\" (UniqueName: \"kubernetes.io/projected/d80ba885-bdec-472c-b9eb-23928eeb6ab1-kube-api-access-d96h2\") pod \"controller-f8648f98b-qq2x8\" (UID: \"d80ba885-bdec-472c-b9eb-23928eeb6ab1\") " pod="metallb-system/controller-f8648f98b-qq2x8" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.106324 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d80ba885-bdec-472c-b9eb-23928eeb6ab1-cert\") pod \"controller-f8648f98b-qq2x8\" (UID: \"d80ba885-bdec-472c-b9eb-23928eeb6ab1\") " pod="metallb-system/controller-f8648f98b-qq2x8" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.108878 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkcpf\" (UniqueName: \"kubernetes.io/projected/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-kube-api-access-zkcpf\") pod \"speaker-g44sf\" (UID: \"e653dfc5-42b0-4ae3-a622-b64b556d3ccc\") " pod="metallb-system/speaker-g44sf" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.388344 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r5wz" event={"ID":"cb278465-e6d8-48b2-bc17-861968bba45e","Type":"ContainerStarted","Data":"e08ab250923f76b9de8c84337b69eaaff7f3aab2a4312f8d4588c0a0dfaa6d68"} Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.482790 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fac17a35-6b82-4397-855c-a9d2b27bcca6-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-wh624\" (UID: \"fac17a35-6b82-4397-855c-a9d2b27bcca6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.487209 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fac17a35-6b82-4397-855c-a9d2b27bcca6-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-wh624\" (UID: \"fac17a35-6b82-4397-855c-a9d2b27bcca6\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.584740 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-memberlist\") pod \"speaker-g44sf\" (UID: \"e653dfc5-42b0-4ae3-a622-b64b556d3ccc\") " pod="metallb-system/speaker-g44sf" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.584832 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d80ba885-bdec-472c-b9eb-23928eeb6ab1-metrics-certs\") pod \"controller-f8648f98b-qq2x8\" (UID: \"d80ba885-bdec-472c-b9eb-23928eeb6ab1\") " pod="metallb-system/controller-f8648f98b-qq2x8" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.584853 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-metrics-certs\") pod \"speaker-g44sf\" (UID: \"e653dfc5-42b0-4ae3-a622-b64b556d3ccc\") " pod="metallb-system/speaker-g44sf" Dec 04 04:09:10 crc kubenswrapper[4806]: E1204 04:09:10.584909 4806 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 04 04:09:10 crc kubenswrapper[4806]: E1204 04:09:10.585015 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-memberlist podName:e653dfc5-42b0-4ae3-a622-b64b556d3ccc nodeName:}" failed. No retries permitted until 2025-12-04 04:09:11.584991465 +0000 UTC m=+866.443504393 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-memberlist") pod "speaker-g44sf" (UID: "e653dfc5-42b0-4ae3-a622-b64b556d3ccc") : secret "metallb-memberlist" not found Dec 04 04:09:10 crc kubenswrapper[4806]: E1204 04:09:10.585120 4806 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 04 04:09:10 crc kubenswrapper[4806]: E1204 04:09:10.585254 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d80ba885-bdec-472c-b9eb-23928eeb6ab1-metrics-certs podName:d80ba885-bdec-472c-b9eb-23928eeb6ab1 nodeName:}" failed. No retries permitted until 2025-12-04 04:09:11.585218443 +0000 UTC m=+866.443731541 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d80ba885-bdec-472c-b9eb-23928eeb6ab1-metrics-certs") pod "controller-f8648f98b-qq2x8" (UID: "d80ba885-bdec-472c-b9eb-23928eeb6ab1") : secret "controller-certs-secret" not found Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.590046 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-metrics-certs\") pod \"speaker-g44sf\" (UID: \"e653dfc5-42b0-4ae3-a622-b64b556d3ccc\") " pod="metallb-system/speaker-g44sf" Dec 04 04:09:10 crc kubenswrapper[4806]: I1204 04:09:10.694668 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624" Dec 04 04:09:11 crc kubenswrapper[4806]: I1204 04:09:11.204388 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624"] Dec 04 04:09:11 crc kubenswrapper[4806]: I1204 04:09:11.397650 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624" event={"ID":"fac17a35-6b82-4397-855c-a9d2b27bcca6","Type":"ContainerStarted","Data":"c95bf284301fde20fb3066381249948b3bba37170b688eefae7d8da6e6d8cd59"} Dec 04 04:09:11 crc kubenswrapper[4806]: I1204 04:09:11.605695 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-memberlist\") pod \"speaker-g44sf\" (UID: \"e653dfc5-42b0-4ae3-a622-b64b556d3ccc\") " pod="metallb-system/speaker-g44sf" Dec 04 04:09:11 crc kubenswrapper[4806]: I1204 04:09:11.605799 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d80ba885-bdec-472c-b9eb-23928eeb6ab1-metrics-certs\") pod \"controller-f8648f98b-qq2x8\" (UID: \"d80ba885-bdec-472c-b9eb-23928eeb6ab1\") " pod="metallb-system/controller-f8648f98b-qq2x8" Dec 04 04:09:11 crc kubenswrapper[4806]: E1204 04:09:11.605878 4806 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 04 04:09:11 crc kubenswrapper[4806]: E1204 04:09:11.605990 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-memberlist podName:e653dfc5-42b0-4ae3-a622-b64b556d3ccc nodeName:}" failed. No retries permitted until 2025-12-04 04:09:13.605964834 +0000 UTC m=+868.464477792 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-memberlist") pod "speaker-g44sf" (UID: "e653dfc5-42b0-4ae3-a622-b64b556d3ccc") : secret "metallb-memberlist" not found Dec 04 04:09:11 crc kubenswrapper[4806]: I1204 04:09:11.614493 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d80ba885-bdec-472c-b9eb-23928eeb6ab1-metrics-certs\") pod \"controller-f8648f98b-qq2x8\" (UID: \"d80ba885-bdec-472c-b9eb-23928eeb6ab1\") " pod="metallb-system/controller-f8648f98b-qq2x8" Dec 04 04:09:11 crc kubenswrapper[4806]: I1204 04:09:11.770058 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-qq2x8" Dec 04 04:09:12 crc kubenswrapper[4806]: W1204 04:09:12.040880 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd80ba885_bdec_472c_b9eb_23928eeb6ab1.slice/crio-30f4efbaa19f1b605aace600b3248b7b9ec52a77d51303e74f05029dcd9df73b WatchSource:0}: Error finding container 30f4efbaa19f1b605aace600b3248b7b9ec52a77d51303e74f05029dcd9df73b: Status 404 returned error can't find the container with id 30f4efbaa19f1b605aace600b3248b7b9ec52a77d51303e74f05029dcd9df73b Dec 04 04:09:12 crc kubenswrapper[4806]: I1204 04:09:12.041593 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-qq2x8"] Dec 04 04:09:12 crc kubenswrapper[4806]: I1204 04:09:12.412226 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-qq2x8" event={"ID":"d80ba885-bdec-472c-b9eb-23928eeb6ab1","Type":"ContainerStarted","Data":"a3819934f313b18fbb6b843bfffab12605275f5f4d1d1cdbd167a2d6b82b9332"} Dec 04 04:09:12 crc kubenswrapper[4806]: I1204 04:09:12.412784 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-qq2x8" event={"ID":"d80ba885-bdec-472c-b9eb-23928eeb6ab1","Type":"ContainerStarted","Data":"2d975060107c41a7cc4887967786b66ae8030a5ce356c11cdbedef0328b9ec73"} Dec 04 04:09:12 crc kubenswrapper[4806]: I1204 04:09:12.412799 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-qq2x8" event={"ID":"d80ba885-bdec-472c-b9eb-23928eeb6ab1","Type":"ContainerStarted","Data":"30f4efbaa19f1b605aace600b3248b7b9ec52a77d51303e74f05029dcd9df73b"} Dec 04 04:09:12 crc kubenswrapper[4806]: I1204 04:09:12.413227 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-qq2x8" Dec 04 04:09:12 crc kubenswrapper[4806]: I1204 04:09:12.453467 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-qq2x8" podStartSLOduration=3.453434016 podStartE2EDuration="3.453434016s" podCreationTimestamp="2025-12-04 04:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:09:12.449947656 +0000 UTC m=+867.308460584" watchObservedRunningTime="2025-12-04 04:09:12.453434016 +0000 UTC m=+867.311946934" Dec 04 04:09:13 crc kubenswrapper[4806]: I1204 04:09:13.649384 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-memberlist\") pod \"speaker-g44sf\" (UID: \"e653dfc5-42b0-4ae3-a622-b64b556d3ccc\") " pod="metallb-system/speaker-g44sf" Dec 04 04:09:13 crc kubenswrapper[4806]: I1204 04:09:13.671069 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e653dfc5-42b0-4ae3-a622-b64b556d3ccc-memberlist\") pod \"speaker-g44sf\" (UID: \"e653dfc5-42b0-4ae3-a622-b64b556d3ccc\") " pod="metallb-system/speaker-g44sf" Dec 04 04:09:13 crc kubenswrapper[4806]: I1204 04:09:13.850390 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-g44sf" Dec 04 04:09:14 crc kubenswrapper[4806]: I1204 04:09:14.442418 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-g44sf" event={"ID":"e653dfc5-42b0-4ae3-a622-b64b556d3ccc","Type":"ContainerStarted","Data":"8254a1cced17b59f3c5c56d93802145c5f9f44a0752e6e78b39b187c82d63ee9"} Dec 04 04:09:15 crc kubenswrapper[4806]: I1204 04:09:15.677630 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-g44sf" podStartSLOduration=6.677597931 podStartE2EDuration="6.677597931s" podCreationTimestamp="2025-12-04 04:09:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:09:15.670619311 +0000 UTC m=+870.529132239" watchObservedRunningTime="2025-12-04 04:09:15.677597931 +0000 UTC m=+870.536110869" Dec 04 04:09:15 crc kubenswrapper[4806]: I1204 04:09:15.749692 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-g44sf" Dec 04 04:09:15 crc kubenswrapper[4806]: I1204 04:09:15.749749 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-g44sf" event={"ID":"e653dfc5-42b0-4ae3-a622-b64b556d3ccc","Type":"ContainerStarted","Data":"3a7263279b105f15d0bcb0debfeee515286fd05d893212d109b2c2a4afa81d07"} Dec 04 04:09:15 crc kubenswrapper[4806]: I1204 04:09:15.749785 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-g44sf" event={"ID":"e653dfc5-42b0-4ae3-a622-b64b556d3ccc","Type":"ContainerStarted","Data":"93671ba4fbecb2e5ea1dda19111da5a020853b47058462b4e29284d893e085cb"} Dec 04 04:09:24 crc kubenswrapper[4806]: I1204 04:09:24.874289 4806 generic.go:334] "Generic (PLEG): container finished" podID="cb278465-e6d8-48b2-bc17-861968bba45e" containerID="e09ed2bd0895b4c6f9d208c27621cc6e98df1ea0e5e40123f1d492609123b0bd" exitCode=0 Dec 04 04:09:24 crc kubenswrapper[4806]: I1204 04:09:24.874576 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r5wz" event={"ID":"cb278465-e6d8-48b2-bc17-861968bba45e","Type":"ContainerDied","Data":"e09ed2bd0895b4c6f9d208c27621cc6e98df1ea0e5e40123f1d492609123b0bd"} Dec 04 04:09:24 crc kubenswrapper[4806]: I1204 04:09:24.877550 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624" event={"ID":"fac17a35-6b82-4397-855c-a9d2b27bcca6","Type":"ContainerStarted","Data":"f8fbed9d2cba629fa24fdb8b70c2545c86d12c2f6230b9eea7c9df35825cab72"} Dec 04 04:09:24 crc kubenswrapper[4806]: I1204 04:09:24.877819 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624" Dec 04 04:09:24 crc kubenswrapper[4806]: I1204 04:09:24.926987 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624" podStartSLOduration=2.50589688 podStartE2EDuration="15.926962518s" podCreationTimestamp="2025-12-04 04:09:09 +0000 UTC" firstStartedPulling="2025-12-04 04:09:11.204239306 +0000 UTC m=+866.062752234" lastFinishedPulling="2025-12-04 04:09:24.625304944 +0000 UTC m=+879.483817872" observedRunningTime="2025-12-04 04:09:24.921143475 +0000 UTC m=+879.779656413" watchObservedRunningTime="2025-12-04 04:09:24.926962518 +0000 UTC m=+879.785475446" Dec 04 04:09:25 crc kubenswrapper[4806]: I1204 04:09:25.888228 4806 generic.go:334] "Generic (PLEG): container finished" podID="cb278465-e6d8-48b2-bc17-861968bba45e" containerID="a748ad7c34e5edbcdc8a52adcefbd1b43f97c6123464995fccefd0f7253b992f" exitCode=0 Dec 04 04:09:25 crc kubenswrapper[4806]: I1204 04:09:25.888321 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r5wz" event={"ID":"cb278465-e6d8-48b2-bc17-861968bba45e","Type":"ContainerDied","Data":"a748ad7c34e5edbcdc8a52adcefbd1b43f97c6123464995fccefd0f7253b992f"} Dec 04 04:09:26 crc kubenswrapper[4806]: I1204 04:09:26.899941 4806 generic.go:334] "Generic (PLEG): container finished" podID="cb278465-e6d8-48b2-bc17-861968bba45e" containerID="d2cd6e8a1b14a6e60e1b15cf987fd19c454325917fd278f773163dd4212694a5" exitCode=0 Dec 04 04:09:26 crc kubenswrapper[4806]: I1204 04:09:26.900492 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r5wz" event={"ID":"cb278465-e6d8-48b2-bc17-861968bba45e","Type":"ContainerDied","Data":"d2cd6e8a1b14a6e60e1b15cf987fd19c454325917fd278f773163dd4212694a5"} Dec 04 04:09:27 crc kubenswrapper[4806]: I1204 04:09:27.047341 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:09:27 crc kubenswrapper[4806]: I1204 04:09:27.047456 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:09:27 crc kubenswrapper[4806]: I1204 04:09:27.913327 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r5wz" event={"ID":"cb278465-e6d8-48b2-bc17-861968bba45e","Type":"ContainerStarted","Data":"2b80abcf821034bc124b49966643f2ce25cc8cac65589b8a2b62fe802315b022"} Dec 04 04:09:27 crc kubenswrapper[4806]: I1204 04:09:27.913643 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r5wz" event={"ID":"cb278465-e6d8-48b2-bc17-861968bba45e","Type":"ContainerStarted","Data":"86ecf4e0a0011fd4f18947a61082bfc194c69157ede5b5b96e10c7c66262dce5"} Dec 04 04:09:27 crc kubenswrapper[4806]: I1204 04:09:27.913653 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r5wz" event={"ID":"cb278465-e6d8-48b2-bc17-861968bba45e","Type":"ContainerStarted","Data":"59c02ab804309fa5eb90c962f85134e98ed0005c231c9b4d1245a3f3e75a0c69"} Dec 04 04:09:27 crc kubenswrapper[4806]: I1204 04:09:27.913662 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r5wz" event={"ID":"cb278465-e6d8-48b2-bc17-861968bba45e","Type":"ContainerStarted","Data":"0cbbbb63b68cde915b889b721052d3ca1519fe4c45851e5cbc91221f8d23cd37"} Dec 04 04:09:27 crc kubenswrapper[4806]: I1204 04:09:27.913670 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r5wz" event={"ID":"cb278465-e6d8-48b2-bc17-861968bba45e","Type":"ContainerStarted","Data":"5238ebd4b24938065335d875172777aa0b676c6718f87814f6fe0ecce47e1b26"} Dec 04 04:09:28 crc kubenswrapper[4806]: I1204 04:09:28.939043 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r5wz" event={"ID":"cb278465-e6d8-48b2-bc17-861968bba45e","Type":"ContainerStarted","Data":"387966177c9eee715c1f1c3767c90484ba756cf53c1493831ea1753551a54f81"} Dec 04 04:09:28 crc kubenswrapper[4806]: I1204 04:09:28.940142 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:28 crc kubenswrapper[4806]: I1204 04:09:28.974306 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-6r5wz" podStartSLOduration=5.673924249 podStartE2EDuration="19.974286271s" podCreationTimestamp="2025-12-04 04:09:09 +0000 UTC" firstStartedPulling="2025-12-04 04:09:10.298908271 +0000 UTC m=+865.157421209" lastFinishedPulling="2025-12-04 04:09:24.599270303 +0000 UTC m=+879.457783231" observedRunningTime="2025-12-04 04:09:28.973680522 +0000 UTC m=+883.832193490" watchObservedRunningTime="2025-12-04 04:09:28.974286271 +0000 UTC m=+883.832799219" Dec 04 04:09:30 crc kubenswrapper[4806]: I1204 04:09:30.057968 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:30 crc kubenswrapper[4806]: I1204 04:09:30.098445 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:31 crc kubenswrapper[4806]: I1204 04:09:31.774069 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-qq2x8" Dec 04 04:09:33 crc kubenswrapper[4806]: I1204 04:09:33.855957 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-g44sf" Dec 04 04:09:36 crc kubenswrapper[4806]: I1204 04:09:36.602981 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-vh5m4"] Dec 04 04:09:36 crc kubenswrapper[4806]: I1204 04:09:36.603975 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vh5m4" Dec 04 04:09:36 crc kubenswrapper[4806]: I1204 04:09:36.610794 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 04 04:09:36 crc kubenswrapper[4806]: I1204 04:09:36.611168 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 04 04:09:36 crc kubenswrapper[4806]: I1204 04:09:36.611992 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-276kj" Dec 04 04:09:36 crc kubenswrapper[4806]: I1204 04:09:36.623311 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vh5m4"] Dec 04 04:09:36 crc kubenswrapper[4806]: I1204 04:09:36.681282 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lllvp\" (UniqueName: \"kubernetes.io/projected/405ec885-56bc-4a69-b7f0-2400da44dc44-kube-api-access-lllvp\") pod \"openstack-operator-index-vh5m4\" (UID: \"405ec885-56bc-4a69-b7f0-2400da44dc44\") " pod="openstack-operators/openstack-operator-index-vh5m4" Dec 04 04:09:36 crc kubenswrapper[4806]: I1204 04:09:36.782378 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lllvp\" (UniqueName: \"kubernetes.io/projected/405ec885-56bc-4a69-b7f0-2400da44dc44-kube-api-access-lllvp\") pod \"openstack-operator-index-vh5m4\" (UID: \"405ec885-56bc-4a69-b7f0-2400da44dc44\") " pod="openstack-operators/openstack-operator-index-vh5m4" Dec 04 04:09:36 crc kubenswrapper[4806]: I1204 04:09:36.810301 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lllvp\" (UniqueName: \"kubernetes.io/projected/405ec885-56bc-4a69-b7f0-2400da44dc44-kube-api-access-lllvp\") pod \"openstack-operator-index-vh5m4\" (UID: \"405ec885-56bc-4a69-b7f0-2400da44dc44\") " pod="openstack-operators/openstack-operator-index-vh5m4" Dec 04 04:09:36 crc kubenswrapper[4806]: I1204 04:09:36.989199 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vh5m4" Dec 04 04:09:37 crc kubenswrapper[4806]: I1204 04:09:37.436121 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vh5m4"] Dec 04 04:09:37 crc kubenswrapper[4806]: W1204 04:09:37.437989 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod405ec885_56bc_4a69_b7f0_2400da44dc44.slice/crio-cbc527b52c0a719bd84c2bfcdc3e25f94387cc2ae5bf38b96f79d5f2c480bdd8 WatchSource:0}: Error finding container cbc527b52c0a719bd84c2bfcdc3e25f94387cc2ae5bf38b96f79d5f2c480bdd8: Status 404 returned error can't find the container with id cbc527b52c0a719bd84c2bfcdc3e25f94387cc2ae5bf38b96f79d5f2c480bdd8 Dec 04 04:09:37 crc kubenswrapper[4806]: I1204 04:09:37.995135 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vh5m4" event={"ID":"405ec885-56bc-4a69-b7f0-2400da44dc44","Type":"ContainerStarted","Data":"cbc527b52c0a719bd84c2bfcdc3e25f94387cc2ae5bf38b96f79d5f2c480bdd8"} Dec 04 04:09:40 crc kubenswrapper[4806]: I1204 04:09:40.060300 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-6r5wz" Dec 04 04:09:40 crc kubenswrapper[4806]: I1204 04:09:40.344309 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-vh5m4"] Dec 04 04:09:40 crc kubenswrapper[4806]: I1204 04:09:40.701255 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-wh624" Dec 04 04:09:40 crc kubenswrapper[4806]: I1204 04:09:40.955330 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-gsvwx"] Dec 04 04:09:40 crc kubenswrapper[4806]: I1204 04:09:40.957041 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-gsvwx" Dec 04 04:09:40 crc kubenswrapper[4806]: I1204 04:09:40.964557 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-gsvwx"] Dec 04 04:09:41 crc kubenswrapper[4806]: I1204 04:09:41.017734 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vh5m4" event={"ID":"405ec885-56bc-4a69-b7f0-2400da44dc44","Type":"ContainerStarted","Data":"55f5e7f09f75b717e1739ab2d5b9901757fb188ad2242aa0e86c29f6fe249155"} Dec 04 04:09:41 crc kubenswrapper[4806]: I1204 04:09:41.047204 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-vh5m4" podStartSLOduration=2.476555397 podStartE2EDuration="5.047181s" podCreationTimestamp="2025-12-04 04:09:36 +0000 UTC" firstStartedPulling="2025-12-04 04:09:37.44613589 +0000 UTC m=+892.304648818" lastFinishedPulling="2025-12-04 04:09:40.016761493 +0000 UTC m=+894.875274421" observedRunningTime="2025-12-04 04:09:41.044261788 +0000 UTC m=+895.902774736" watchObservedRunningTime="2025-12-04 04:09:41.047181 +0000 UTC m=+895.905693948" Dec 04 04:09:41 crc kubenswrapper[4806]: I1204 04:09:41.092409 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-487xh\" (UniqueName: \"kubernetes.io/projected/89d5c50c-5d00-4f21-9e60-8c6692a0ad33-kube-api-access-487xh\") pod \"openstack-operator-index-gsvwx\" (UID: \"89d5c50c-5d00-4f21-9e60-8c6692a0ad33\") " pod="openstack-operators/openstack-operator-index-gsvwx" Dec 04 04:09:41 crc kubenswrapper[4806]: I1204 04:09:41.193610 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-487xh\" (UniqueName: \"kubernetes.io/projected/89d5c50c-5d00-4f21-9e60-8c6692a0ad33-kube-api-access-487xh\") pod \"openstack-operator-index-gsvwx\" (UID: \"89d5c50c-5d00-4f21-9e60-8c6692a0ad33\") " pod="openstack-operators/openstack-operator-index-gsvwx" Dec 04 04:09:41 crc kubenswrapper[4806]: I1204 04:09:41.220891 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-487xh\" (UniqueName: \"kubernetes.io/projected/89d5c50c-5d00-4f21-9e60-8c6692a0ad33-kube-api-access-487xh\") pod \"openstack-operator-index-gsvwx\" (UID: \"89d5c50c-5d00-4f21-9e60-8c6692a0ad33\") " pod="openstack-operators/openstack-operator-index-gsvwx" Dec 04 04:09:41 crc kubenswrapper[4806]: I1204 04:09:41.288832 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-gsvwx" Dec 04 04:09:41 crc kubenswrapper[4806]: I1204 04:09:41.480963 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-gsvwx"] Dec 04 04:09:41 crc kubenswrapper[4806]: W1204 04:09:41.488975 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89d5c50c_5d00_4f21_9e60_8c6692a0ad33.slice/crio-cbcb64e869efa463ecbaba61c3a2cc10035d8e44a5632a36ef03a0a4887cfe69 WatchSource:0}: Error finding container cbcb64e869efa463ecbaba61c3a2cc10035d8e44a5632a36ef03a0a4887cfe69: Status 404 returned error can't find the container with id cbcb64e869efa463ecbaba61c3a2cc10035d8e44a5632a36ef03a0a4887cfe69 Dec 04 04:09:42 crc kubenswrapper[4806]: I1204 04:09:42.024054 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-vh5m4" podUID="405ec885-56bc-4a69-b7f0-2400da44dc44" containerName="registry-server" containerID="cri-o://55f5e7f09f75b717e1739ab2d5b9901757fb188ad2242aa0e86c29f6fe249155" gracePeriod=2 Dec 04 04:09:42 crc kubenswrapper[4806]: I1204 04:09:42.024113 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-gsvwx" event={"ID":"89d5c50c-5d00-4f21-9e60-8c6692a0ad33","Type":"ContainerStarted","Data":"cbcb64e869efa463ecbaba61c3a2cc10035d8e44a5632a36ef03a0a4887cfe69"} Dec 04 04:09:42 crc kubenswrapper[4806]: I1204 04:09:42.445625 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vh5m4" Dec 04 04:09:42 crc kubenswrapper[4806]: I1204 04:09:42.511744 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lllvp\" (UniqueName: \"kubernetes.io/projected/405ec885-56bc-4a69-b7f0-2400da44dc44-kube-api-access-lllvp\") pod \"405ec885-56bc-4a69-b7f0-2400da44dc44\" (UID: \"405ec885-56bc-4a69-b7f0-2400da44dc44\") " Dec 04 04:09:42 crc kubenswrapper[4806]: I1204 04:09:42.518203 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/405ec885-56bc-4a69-b7f0-2400da44dc44-kube-api-access-lllvp" (OuterVolumeSpecName: "kube-api-access-lllvp") pod "405ec885-56bc-4a69-b7f0-2400da44dc44" (UID: "405ec885-56bc-4a69-b7f0-2400da44dc44"). InnerVolumeSpecName "kube-api-access-lllvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:09:42 crc kubenswrapper[4806]: I1204 04:09:42.613709 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lllvp\" (UniqueName: \"kubernetes.io/projected/405ec885-56bc-4a69-b7f0-2400da44dc44-kube-api-access-lllvp\") on node \"crc\" DevicePath \"\"" Dec 04 04:09:43 crc kubenswrapper[4806]: I1204 04:09:43.030505 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-gsvwx" event={"ID":"89d5c50c-5d00-4f21-9e60-8c6692a0ad33","Type":"ContainerStarted","Data":"7c8410605f5f25e107b7c57b2c99f4d1359962a0aaf5364134cf4e4cce0fbc72"} Dec 04 04:09:43 crc kubenswrapper[4806]: I1204 04:09:43.032396 4806 generic.go:334] "Generic (PLEG): container finished" podID="405ec885-56bc-4a69-b7f0-2400da44dc44" containerID="55f5e7f09f75b717e1739ab2d5b9901757fb188ad2242aa0e86c29f6fe249155" exitCode=0 Dec 04 04:09:43 crc kubenswrapper[4806]: I1204 04:09:43.032440 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vh5m4" event={"ID":"405ec885-56bc-4a69-b7f0-2400da44dc44","Type":"ContainerDied","Data":"55f5e7f09f75b717e1739ab2d5b9901757fb188ad2242aa0e86c29f6fe249155"} Dec 04 04:09:43 crc kubenswrapper[4806]: I1204 04:09:43.032462 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vh5m4" event={"ID":"405ec885-56bc-4a69-b7f0-2400da44dc44","Type":"ContainerDied","Data":"cbc527b52c0a719bd84c2bfcdc3e25f94387cc2ae5bf38b96f79d5f2c480bdd8"} Dec 04 04:09:43 crc kubenswrapper[4806]: I1204 04:09:43.032479 4806 scope.go:117] "RemoveContainer" containerID="55f5e7f09f75b717e1739ab2d5b9901757fb188ad2242aa0e86c29f6fe249155" Dec 04 04:09:43 crc kubenswrapper[4806]: I1204 04:09:43.032572 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vh5m4" Dec 04 04:09:43 crc kubenswrapper[4806]: I1204 04:09:43.045455 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-gsvwx" podStartSLOduration=2.539745636 podStartE2EDuration="3.04543377s" podCreationTimestamp="2025-12-04 04:09:40 +0000 UTC" firstStartedPulling="2025-12-04 04:09:41.492357186 +0000 UTC m=+896.350870114" lastFinishedPulling="2025-12-04 04:09:41.99804532 +0000 UTC m=+896.856558248" observedRunningTime="2025-12-04 04:09:43.042733665 +0000 UTC m=+897.901246593" watchObservedRunningTime="2025-12-04 04:09:43.04543377 +0000 UTC m=+897.903946698" Dec 04 04:09:43 crc kubenswrapper[4806]: I1204 04:09:43.052549 4806 scope.go:117] "RemoveContainer" containerID="55f5e7f09f75b717e1739ab2d5b9901757fb188ad2242aa0e86c29f6fe249155" Dec 04 04:09:43 crc kubenswrapper[4806]: E1204 04:09:43.052957 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55f5e7f09f75b717e1739ab2d5b9901757fb188ad2242aa0e86c29f6fe249155\": container with ID starting with 55f5e7f09f75b717e1739ab2d5b9901757fb188ad2242aa0e86c29f6fe249155 not found: ID does not exist" containerID="55f5e7f09f75b717e1739ab2d5b9901757fb188ad2242aa0e86c29f6fe249155" Dec 04 04:09:43 crc kubenswrapper[4806]: I1204 04:09:43.052998 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55f5e7f09f75b717e1739ab2d5b9901757fb188ad2242aa0e86c29f6fe249155"} err="failed to get container status \"55f5e7f09f75b717e1739ab2d5b9901757fb188ad2242aa0e86c29f6fe249155\": rpc error: code = NotFound desc = could not find container \"55f5e7f09f75b717e1739ab2d5b9901757fb188ad2242aa0e86c29f6fe249155\": container with ID starting with 55f5e7f09f75b717e1739ab2d5b9901757fb188ad2242aa0e86c29f6fe249155 not found: ID does not exist" Dec 04 04:09:43 crc kubenswrapper[4806]: I1204 04:09:43.072519 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-vh5m4"] Dec 04 04:09:43 crc kubenswrapper[4806]: I1204 04:09:43.076601 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-vh5m4"] Dec 04 04:09:43 crc kubenswrapper[4806]: I1204 04:09:43.435177 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="405ec885-56bc-4a69-b7f0-2400da44dc44" path="/var/lib/kubelet/pods/405ec885-56bc-4a69-b7f0-2400da44dc44/volumes" Dec 04 04:09:47 crc kubenswrapper[4806]: I1204 04:09:47.949804 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p8gtf"] Dec 04 04:09:47 crc kubenswrapper[4806]: E1204 04:09:47.950991 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="405ec885-56bc-4a69-b7f0-2400da44dc44" containerName="registry-server" Dec 04 04:09:47 crc kubenswrapper[4806]: I1204 04:09:47.951012 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="405ec885-56bc-4a69-b7f0-2400da44dc44" containerName="registry-server" Dec 04 04:09:47 crc kubenswrapper[4806]: I1204 04:09:47.951163 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="405ec885-56bc-4a69-b7f0-2400da44dc44" containerName="registry-server" Dec 04 04:09:47 crc kubenswrapper[4806]: I1204 04:09:47.952036 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:09:47 crc kubenswrapper[4806]: I1204 04:09:47.965418 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p8gtf"] Dec 04 04:09:47 crc kubenswrapper[4806]: I1204 04:09:47.994138 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73119d3c-cd1e-48d4-bb3b-ea456876feb3-utilities\") pod \"certified-operators-p8gtf\" (UID: \"73119d3c-cd1e-48d4-bb3b-ea456876feb3\") " pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:09:47 crc kubenswrapper[4806]: I1204 04:09:47.994497 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73119d3c-cd1e-48d4-bb3b-ea456876feb3-catalog-content\") pod \"certified-operators-p8gtf\" (UID: \"73119d3c-cd1e-48d4-bb3b-ea456876feb3\") " pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:09:47 crc kubenswrapper[4806]: I1204 04:09:47.994608 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzscq\" (UniqueName: \"kubernetes.io/projected/73119d3c-cd1e-48d4-bb3b-ea456876feb3-kube-api-access-xzscq\") pod \"certified-operators-p8gtf\" (UID: \"73119d3c-cd1e-48d4-bb3b-ea456876feb3\") " pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:09:48 crc kubenswrapper[4806]: I1204 04:09:48.096520 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73119d3c-cd1e-48d4-bb3b-ea456876feb3-catalog-content\") pod \"certified-operators-p8gtf\" (UID: \"73119d3c-cd1e-48d4-bb3b-ea456876feb3\") " pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:09:48 crc kubenswrapper[4806]: I1204 04:09:48.096566 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzscq\" (UniqueName: \"kubernetes.io/projected/73119d3c-cd1e-48d4-bb3b-ea456876feb3-kube-api-access-xzscq\") pod \"certified-operators-p8gtf\" (UID: \"73119d3c-cd1e-48d4-bb3b-ea456876feb3\") " pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:09:48 crc kubenswrapper[4806]: I1204 04:09:48.096637 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73119d3c-cd1e-48d4-bb3b-ea456876feb3-utilities\") pod \"certified-operators-p8gtf\" (UID: \"73119d3c-cd1e-48d4-bb3b-ea456876feb3\") " pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:09:48 crc kubenswrapper[4806]: I1204 04:09:48.097102 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73119d3c-cd1e-48d4-bb3b-ea456876feb3-catalog-content\") pod \"certified-operators-p8gtf\" (UID: \"73119d3c-cd1e-48d4-bb3b-ea456876feb3\") " pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:09:48 crc kubenswrapper[4806]: I1204 04:09:48.097166 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73119d3c-cd1e-48d4-bb3b-ea456876feb3-utilities\") pod \"certified-operators-p8gtf\" (UID: \"73119d3c-cd1e-48d4-bb3b-ea456876feb3\") " pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:09:48 crc kubenswrapper[4806]: I1204 04:09:48.117070 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzscq\" (UniqueName: \"kubernetes.io/projected/73119d3c-cd1e-48d4-bb3b-ea456876feb3-kube-api-access-xzscq\") pod \"certified-operators-p8gtf\" (UID: \"73119d3c-cd1e-48d4-bb3b-ea456876feb3\") " pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:09:48 crc kubenswrapper[4806]: I1204 04:09:48.303637 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:09:48 crc kubenswrapper[4806]: I1204 04:09:48.743692 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p8gtf"] Dec 04 04:09:49 crc kubenswrapper[4806]: I1204 04:09:49.122662 4806 generic.go:334] "Generic (PLEG): container finished" podID="73119d3c-cd1e-48d4-bb3b-ea456876feb3" containerID="4d6c6c849ac31c895a15b6348140ef4864d0aa4139a49ee8f1231f54bda03ccb" exitCode=0 Dec 04 04:09:49 crc kubenswrapper[4806]: I1204 04:09:49.122707 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8gtf" event={"ID":"73119d3c-cd1e-48d4-bb3b-ea456876feb3","Type":"ContainerDied","Data":"4d6c6c849ac31c895a15b6348140ef4864d0aa4139a49ee8f1231f54bda03ccb"} Dec 04 04:09:49 crc kubenswrapper[4806]: I1204 04:09:49.122737 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8gtf" event={"ID":"73119d3c-cd1e-48d4-bb3b-ea456876feb3","Type":"ContainerStarted","Data":"bb3e33666b334816b417997efe6bb51e57f79609882cd5199b9158eb5b35d6a3"} Dec 04 04:09:50 crc kubenswrapper[4806]: I1204 04:09:50.133915 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8gtf" event={"ID":"73119d3c-cd1e-48d4-bb3b-ea456876feb3","Type":"ContainerStarted","Data":"6bfa4189975313c0da965e4ef3386b9a14a4693dde6f4dae524f6b02327b0605"} Dec 04 04:09:51 crc kubenswrapper[4806]: I1204 04:09:51.144019 4806 generic.go:334] "Generic (PLEG): container finished" podID="73119d3c-cd1e-48d4-bb3b-ea456876feb3" containerID="6bfa4189975313c0da965e4ef3386b9a14a4693dde6f4dae524f6b02327b0605" exitCode=0 Dec 04 04:09:51 crc kubenswrapper[4806]: I1204 04:09:51.144074 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8gtf" event={"ID":"73119d3c-cd1e-48d4-bb3b-ea456876feb3","Type":"ContainerDied","Data":"6bfa4189975313c0da965e4ef3386b9a14a4693dde6f4dae524f6b02327b0605"} Dec 04 04:09:51 crc kubenswrapper[4806]: I1204 04:09:51.289529 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-gsvwx" Dec 04 04:09:51 crc kubenswrapper[4806]: I1204 04:09:51.289608 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-gsvwx" Dec 04 04:09:51 crc kubenswrapper[4806]: I1204 04:09:51.317749 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-gsvwx" Dec 04 04:09:52 crc kubenswrapper[4806]: I1204 04:09:52.151837 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8gtf" event={"ID":"73119d3c-cd1e-48d4-bb3b-ea456876feb3","Type":"ContainerStarted","Data":"60fa51535ae47c0cd7751def60e4b7272b1ba9563d27b73c17ba1ae663082cdf"} Dec 04 04:09:52 crc kubenswrapper[4806]: I1204 04:09:52.174698 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p8gtf" podStartSLOduration=2.6880505230000002 podStartE2EDuration="5.17468166s" podCreationTimestamp="2025-12-04 04:09:47 +0000 UTC" firstStartedPulling="2025-12-04 04:09:49.125533219 +0000 UTC m=+903.984046147" lastFinishedPulling="2025-12-04 04:09:51.612164316 +0000 UTC m=+906.470677284" observedRunningTime="2025-12-04 04:09:52.172696887 +0000 UTC m=+907.031209815" watchObservedRunningTime="2025-12-04 04:09:52.17468166 +0000 UTC m=+907.033194588" Dec 04 04:09:52 crc kubenswrapper[4806]: I1204 04:09:52.182361 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-gsvwx" Dec 04 04:09:57 crc kubenswrapper[4806]: I1204 04:09:57.047174 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:09:57 crc kubenswrapper[4806]: I1204 04:09:57.047685 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:09:58 crc kubenswrapper[4806]: I1204 04:09:58.304767 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:09:58 crc kubenswrapper[4806]: I1204 04:09:58.304813 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:09:58 crc kubenswrapper[4806]: I1204 04:09:58.475899 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:09:59 crc kubenswrapper[4806]: I1204 04:09:59.269513 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:09:59 crc kubenswrapper[4806]: I1204 04:09:59.938207 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p8gtf"] Dec 04 04:10:00 crc kubenswrapper[4806]: I1204 04:10:00.187278 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh"] Dec 04 04:10:00 crc kubenswrapper[4806]: I1204 04:10:00.188891 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" Dec 04 04:10:00 crc kubenswrapper[4806]: I1204 04:10:00.190981 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-tdwxq" Dec 04 04:10:00 crc kubenswrapper[4806]: I1204 04:10:00.203181 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh"] Dec 04 04:10:00 crc kubenswrapper[4806]: I1204 04:10:00.384806 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a72e0bcd-e604-4243-b294-70ebe7ba990c-util\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh\" (UID: \"a72e0bcd-e604-4243-b294-70ebe7ba990c\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" Dec 04 04:10:00 crc kubenswrapper[4806]: I1204 04:10:00.385534 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbxqf\" (UniqueName: \"kubernetes.io/projected/a72e0bcd-e604-4243-b294-70ebe7ba990c-kube-api-access-zbxqf\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh\" (UID: \"a72e0bcd-e604-4243-b294-70ebe7ba990c\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" Dec 04 04:10:00 crc kubenswrapper[4806]: I1204 04:10:00.385626 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a72e0bcd-e604-4243-b294-70ebe7ba990c-bundle\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh\" (UID: \"a72e0bcd-e604-4243-b294-70ebe7ba990c\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" Dec 04 04:10:00 crc kubenswrapper[4806]: I1204 04:10:00.486842 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbxqf\" (UniqueName: \"kubernetes.io/projected/a72e0bcd-e604-4243-b294-70ebe7ba990c-kube-api-access-zbxqf\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh\" (UID: \"a72e0bcd-e604-4243-b294-70ebe7ba990c\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" Dec 04 04:10:00 crc kubenswrapper[4806]: I1204 04:10:00.486913 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a72e0bcd-e604-4243-b294-70ebe7ba990c-bundle\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh\" (UID: \"a72e0bcd-e604-4243-b294-70ebe7ba990c\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" Dec 04 04:10:00 crc kubenswrapper[4806]: I1204 04:10:00.486960 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a72e0bcd-e604-4243-b294-70ebe7ba990c-util\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh\" (UID: \"a72e0bcd-e604-4243-b294-70ebe7ba990c\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" Dec 04 04:10:00 crc kubenswrapper[4806]: I1204 04:10:00.487471 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a72e0bcd-e604-4243-b294-70ebe7ba990c-util\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh\" (UID: \"a72e0bcd-e604-4243-b294-70ebe7ba990c\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" Dec 04 04:10:00 crc kubenswrapper[4806]: I1204 04:10:00.487571 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a72e0bcd-e604-4243-b294-70ebe7ba990c-bundle\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh\" (UID: \"a72e0bcd-e604-4243-b294-70ebe7ba990c\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" Dec 04 04:10:00 crc kubenswrapper[4806]: I1204 04:10:00.508497 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbxqf\" (UniqueName: \"kubernetes.io/projected/a72e0bcd-e604-4243-b294-70ebe7ba990c-kube-api-access-zbxqf\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh\" (UID: \"a72e0bcd-e604-4243-b294-70ebe7ba990c\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" Dec 04 04:10:00 crc kubenswrapper[4806]: I1204 04:10:00.566798 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" Dec 04 04:10:01 crc kubenswrapper[4806]: I1204 04:10:01.070790 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh"] Dec 04 04:10:01 crc kubenswrapper[4806]: W1204 04:10:01.092110 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda72e0bcd_e604_4243_b294_70ebe7ba990c.slice/crio-e19410f73dc2aaf9580117324e55d206e3154b388ac07298b22c8a7ddea81a75 WatchSource:0}: Error finding container e19410f73dc2aaf9580117324e55d206e3154b388ac07298b22c8a7ddea81a75: Status 404 returned error can't find the container with id e19410f73dc2aaf9580117324e55d206e3154b388ac07298b22c8a7ddea81a75 Dec 04 04:10:01 crc kubenswrapper[4806]: I1204 04:10:01.260375 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" event={"ID":"a72e0bcd-e604-4243-b294-70ebe7ba990c","Type":"ContainerStarted","Data":"e19410f73dc2aaf9580117324e55d206e3154b388ac07298b22c8a7ddea81a75"} Dec 04 04:10:01 crc kubenswrapper[4806]: I1204 04:10:01.260572 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p8gtf" podUID="73119d3c-cd1e-48d4-bb3b-ea456876feb3" containerName="registry-server" containerID="cri-o://60fa51535ae47c0cd7751def60e4b7272b1ba9563d27b73c17ba1ae663082cdf" gracePeriod=2 Dec 04 04:10:02 crc kubenswrapper[4806]: I1204 04:10:02.267025 4806 generic.go:334] "Generic (PLEG): container finished" podID="a72e0bcd-e604-4243-b294-70ebe7ba990c" containerID="695dc80ad0c7b0d86a70de02219f14b67d8221a5e5110e4a80aa4abeabbf6984" exitCode=0 Dec 04 04:10:02 crc kubenswrapper[4806]: I1204 04:10:02.267091 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" event={"ID":"a72e0bcd-e604-4243-b294-70ebe7ba990c","Type":"ContainerDied","Data":"695dc80ad0c7b0d86a70de02219f14b67d8221a5e5110e4a80aa4abeabbf6984"} Dec 04 04:10:02 crc kubenswrapper[4806]: I1204 04:10:02.274098 4806 generic.go:334] "Generic (PLEG): container finished" podID="73119d3c-cd1e-48d4-bb3b-ea456876feb3" containerID="60fa51535ae47c0cd7751def60e4b7272b1ba9563d27b73c17ba1ae663082cdf" exitCode=0 Dec 04 04:10:02 crc kubenswrapper[4806]: I1204 04:10:02.274151 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8gtf" event={"ID":"73119d3c-cd1e-48d4-bb3b-ea456876feb3","Type":"ContainerDied","Data":"60fa51535ae47c0cd7751def60e4b7272b1ba9563d27b73c17ba1ae663082cdf"} Dec 04 04:10:02 crc kubenswrapper[4806]: I1204 04:10:02.528805 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:10:02 crc kubenswrapper[4806]: I1204 04:10:02.721159 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzscq\" (UniqueName: \"kubernetes.io/projected/73119d3c-cd1e-48d4-bb3b-ea456876feb3-kube-api-access-xzscq\") pod \"73119d3c-cd1e-48d4-bb3b-ea456876feb3\" (UID: \"73119d3c-cd1e-48d4-bb3b-ea456876feb3\") " Dec 04 04:10:02 crc kubenswrapper[4806]: I1204 04:10:02.721253 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73119d3c-cd1e-48d4-bb3b-ea456876feb3-utilities\") pod \"73119d3c-cd1e-48d4-bb3b-ea456876feb3\" (UID: \"73119d3c-cd1e-48d4-bb3b-ea456876feb3\") " Dec 04 04:10:02 crc kubenswrapper[4806]: I1204 04:10:02.721413 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73119d3c-cd1e-48d4-bb3b-ea456876feb3-catalog-content\") pod \"73119d3c-cd1e-48d4-bb3b-ea456876feb3\" (UID: \"73119d3c-cd1e-48d4-bb3b-ea456876feb3\") " Dec 04 04:10:02 crc kubenswrapper[4806]: I1204 04:10:02.722582 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73119d3c-cd1e-48d4-bb3b-ea456876feb3-utilities" (OuterVolumeSpecName: "utilities") pod "73119d3c-cd1e-48d4-bb3b-ea456876feb3" (UID: "73119d3c-cd1e-48d4-bb3b-ea456876feb3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:10:02 crc kubenswrapper[4806]: I1204 04:10:02.743433 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73119d3c-cd1e-48d4-bb3b-ea456876feb3-kube-api-access-xzscq" (OuterVolumeSpecName: "kube-api-access-xzscq") pod "73119d3c-cd1e-48d4-bb3b-ea456876feb3" (UID: "73119d3c-cd1e-48d4-bb3b-ea456876feb3"). InnerVolumeSpecName "kube-api-access-xzscq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:10:02 crc kubenswrapper[4806]: I1204 04:10:02.781292 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73119d3c-cd1e-48d4-bb3b-ea456876feb3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73119d3c-cd1e-48d4-bb3b-ea456876feb3" (UID: "73119d3c-cd1e-48d4-bb3b-ea456876feb3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:10:02 crc kubenswrapper[4806]: I1204 04:10:02.823281 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73119d3c-cd1e-48d4-bb3b-ea456876feb3-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:10:02 crc kubenswrapper[4806]: I1204 04:10:02.823326 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73119d3c-cd1e-48d4-bb3b-ea456876feb3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:10:02 crc kubenswrapper[4806]: I1204 04:10:02.823345 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzscq\" (UniqueName: \"kubernetes.io/projected/73119d3c-cd1e-48d4-bb3b-ea456876feb3-kube-api-access-xzscq\") on node \"crc\" DevicePath \"\"" Dec 04 04:10:03 crc kubenswrapper[4806]: I1204 04:10:03.282195 4806 generic.go:334] "Generic (PLEG): container finished" podID="a72e0bcd-e604-4243-b294-70ebe7ba990c" containerID="db311037da7df2b622b37bba26696b50650998a1317013aa30bb13c016754375" exitCode=0 Dec 04 04:10:03 crc kubenswrapper[4806]: I1204 04:10:03.282287 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" event={"ID":"a72e0bcd-e604-4243-b294-70ebe7ba990c","Type":"ContainerDied","Data":"db311037da7df2b622b37bba26696b50650998a1317013aa30bb13c016754375"} Dec 04 04:10:03 crc kubenswrapper[4806]: I1204 04:10:03.285693 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8gtf" event={"ID":"73119d3c-cd1e-48d4-bb3b-ea456876feb3","Type":"ContainerDied","Data":"bb3e33666b334816b417997efe6bb51e57f79609882cd5199b9158eb5b35d6a3"} Dec 04 04:10:03 crc kubenswrapper[4806]: I1204 04:10:03.285746 4806 scope.go:117] "RemoveContainer" containerID="60fa51535ae47c0cd7751def60e4b7272b1ba9563d27b73c17ba1ae663082cdf" Dec 04 04:10:03 crc kubenswrapper[4806]: I1204 04:10:03.285778 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p8gtf" Dec 04 04:10:03 crc kubenswrapper[4806]: I1204 04:10:03.311015 4806 scope.go:117] "RemoveContainer" containerID="6bfa4189975313c0da965e4ef3386b9a14a4693dde6f4dae524f6b02327b0605" Dec 04 04:10:03 crc kubenswrapper[4806]: I1204 04:10:03.345432 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p8gtf"] Dec 04 04:10:03 crc kubenswrapper[4806]: I1204 04:10:03.354687 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p8gtf"] Dec 04 04:10:03 crc kubenswrapper[4806]: I1204 04:10:03.354900 4806 scope.go:117] "RemoveContainer" containerID="4d6c6c849ac31c895a15b6348140ef4864d0aa4139a49ee8f1231f54bda03ccb" Dec 04 04:10:03 crc kubenswrapper[4806]: I1204 04:10:03.431642 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73119d3c-cd1e-48d4-bb3b-ea456876feb3" path="/var/lib/kubelet/pods/73119d3c-cd1e-48d4-bb3b-ea456876feb3/volumes" Dec 04 04:10:04 crc kubenswrapper[4806]: I1204 04:10:04.295339 4806 generic.go:334] "Generic (PLEG): container finished" podID="a72e0bcd-e604-4243-b294-70ebe7ba990c" containerID="a71ebf08d9068f704ec4b61069f0d0561a159fbc76a5cd99f283754163c8c686" exitCode=0 Dec 04 04:10:04 crc kubenswrapper[4806]: I1204 04:10:04.295527 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" event={"ID":"a72e0bcd-e604-4243-b294-70ebe7ba990c","Type":"ContainerDied","Data":"a71ebf08d9068f704ec4b61069f0d0561a159fbc76a5cd99f283754163c8c686"} Dec 04 04:10:05 crc kubenswrapper[4806]: I1204 04:10:05.610767 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" Dec 04 04:10:05 crc kubenswrapper[4806]: I1204 04:10:05.781132 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a72e0bcd-e604-4243-b294-70ebe7ba990c-bundle\") pod \"a72e0bcd-e604-4243-b294-70ebe7ba990c\" (UID: \"a72e0bcd-e604-4243-b294-70ebe7ba990c\") " Dec 04 04:10:05 crc kubenswrapper[4806]: I1204 04:10:05.781215 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbxqf\" (UniqueName: \"kubernetes.io/projected/a72e0bcd-e604-4243-b294-70ebe7ba990c-kube-api-access-zbxqf\") pod \"a72e0bcd-e604-4243-b294-70ebe7ba990c\" (UID: \"a72e0bcd-e604-4243-b294-70ebe7ba990c\") " Dec 04 04:10:05 crc kubenswrapper[4806]: I1204 04:10:05.781308 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a72e0bcd-e604-4243-b294-70ebe7ba990c-util\") pod \"a72e0bcd-e604-4243-b294-70ebe7ba990c\" (UID: \"a72e0bcd-e604-4243-b294-70ebe7ba990c\") " Dec 04 04:10:05 crc kubenswrapper[4806]: I1204 04:10:05.782532 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a72e0bcd-e604-4243-b294-70ebe7ba990c-bundle" (OuterVolumeSpecName: "bundle") pod "a72e0bcd-e604-4243-b294-70ebe7ba990c" (UID: "a72e0bcd-e604-4243-b294-70ebe7ba990c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:10:05 crc kubenswrapper[4806]: I1204 04:10:05.790183 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a72e0bcd-e604-4243-b294-70ebe7ba990c-kube-api-access-zbxqf" (OuterVolumeSpecName: "kube-api-access-zbxqf") pod "a72e0bcd-e604-4243-b294-70ebe7ba990c" (UID: "a72e0bcd-e604-4243-b294-70ebe7ba990c"). InnerVolumeSpecName "kube-api-access-zbxqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:10:05 crc kubenswrapper[4806]: I1204 04:10:05.800142 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a72e0bcd-e604-4243-b294-70ebe7ba990c-util" (OuterVolumeSpecName: "util") pod "a72e0bcd-e604-4243-b294-70ebe7ba990c" (UID: "a72e0bcd-e604-4243-b294-70ebe7ba990c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:10:05 crc kubenswrapper[4806]: I1204 04:10:05.882352 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbxqf\" (UniqueName: \"kubernetes.io/projected/a72e0bcd-e604-4243-b294-70ebe7ba990c-kube-api-access-zbxqf\") on node \"crc\" DevicePath \"\"" Dec 04 04:10:05 crc kubenswrapper[4806]: I1204 04:10:05.882386 4806 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a72e0bcd-e604-4243-b294-70ebe7ba990c-util\") on node \"crc\" DevicePath \"\"" Dec 04 04:10:05 crc kubenswrapper[4806]: I1204 04:10:05.882396 4806 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a72e0bcd-e604-4243-b294-70ebe7ba990c-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:10:06 crc kubenswrapper[4806]: I1204 04:10:06.309582 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" event={"ID":"a72e0bcd-e604-4243-b294-70ebe7ba990c","Type":"ContainerDied","Data":"e19410f73dc2aaf9580117324e55d206e3154b388ac07298b22c8a7ddea81a75"} Dec 04 04:10:06 crc kubenswrapper[4806]: I1204 04:10:06.309627 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e19410f73dc2aaf9580117324e55d206e3154b388ac07298b22c8a7ddea81a75" Dec 04 04:10:06 crc kubenswrapper[4806]: I1204 04:10:06.309652 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh" Dec 04 04:10:09 crc kubenswrapper[4806]: I1204 04:10:09.668100 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7bfcb77fcc-bx765"] Dec 04 04:10:09 crc kubenswrapper[4806]: E1204 04:10:09.668706 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a72e0bcd-e604-4243-b294-70ebe7ba990c" containerName="extract" Dec 04 04:10:09 crc kubenswrapper[4806]: I1204 04:10:09.668724 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a72e0bcd-e604-4243-b294-70ebe7ba990c" containerName="extract" Dec 04 04:10:09 crc kubenswrapper[4806]: E1204 04:10:09.668739 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73119d3c-cd1e-48d4-bb3b-ea456876feb3" containerName="extract-content" Dec 04 04:10:09 crc kubenswrapper[4806]: I1204 04:10:09.668747 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="73119d3c-cd1e-48d4-bb3b-ea456876feb3" containerName="extract-content" Dec 04 04:10:09 crc kubenswrapper[4806]: E1204 04:10:09.668758 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73119d3c-cd1e-48d4-bb3b-ea456876feb3" containerName="registry-server" Dec 04 04:10:09 crc kubenswrapper[4806]: I1204 04:10:09.668769 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="73119d3c-cd1e-48d4-bb3b-ea456876feb3" containerName="registry-server" Dec 04 04:10:09 crc kubenswrapper[4806]: E1204 04:10:09.668779 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a72e0bcd-e604-4243-b294-70ebe7ba990c" containerName="pull" Dec 04 04:10:09 crc kubenswrapper[4806]: I1204 04:10:09.668788 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a72e0bcd-e604-4243-b294-70ebe7ba990c" containerName="pull" Dec 04 04:10:09 crc kubenswrapper[4806]: E1204 04:10:09.668800 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73119d3c-cd1e-48d4-bb3b-ea456876feb3" containerName="extract-utilities" Dec 04 04:10:09 crc kubenswrapper[4806]: I1204 04:10:09.668808 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="73119d3c-cd1e-48d4-bb3b-ea456876feb3" containerName="extract-utilities" Dec 04 04:10:09 crc kubenswrapper[4806]: E1204 04:10:09.668822 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a72e0bcd-e604-4243-b294-70ebe7ba990c" containerName="util" Dec 04 04:10:09 crc kubenswrapper[4806]: I1204 04:10:09.668830 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a72e0bcd-e604-4243-b294-70ebe7ba990c" containerName="util" Dec 04 04:10:09 crc kubenswrapper[4806]: I1204 04:10:09.669050 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="a72e0bcd-e604-4243-b294-70ebe7ba990c" containerName="extract" Dec 04 04:10:09 crc kubenswrapper[4806]: I1204 04:10:09.669068 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="73119d3c-cd1e-48d4-bb3b-ea456876feb3" containerName="registry-server" Dec 04 04:10:09 crc kubenswrapper[4806]: I1204 04:10:09.669496 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7bfcb77fcc-bx765" Dec 04 04:10:09 crc kubenswrapper[4806]: I1204 04:10:09.671983 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-vqjjn" Dec 04 04:10:09 crc kubenswrapper[4806]: I1204 04:10:09.751573 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7bfcb77fcc-bx765"] Dec 04 04:10:09 crc kubenswrapper[4806]: I1204 04:10:09.832810 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h5d4\" (UniqueName: \"kubernetes.io/projected/4d229897-71a8-4b61-a9be-4d963e1384ac-kube-api-access-8h5d4\") pod \"openstack-operator-controller-operator-7bfcb77fcc-bx765\" (UID: \"4d229897-71a8-4b61-a9be-4d963e1384ac\") " pod="openstack-operators/openstack-operator-controller-operator-7bfcb77fcc-bx765" Dec 04 04:10:09 crc kubenswrapper[4806]: I1204 04:10:09.933473 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h5d4\" (UniqueName: \"kubernetes.io/projected/4d229897-71a8-4b61-a9be-4d963e1384ac-kube-api-access-8h5d4\") pod \"openstack-operator-controller-operator-7bfcb77fcc-bx765\" (UID: \"4d229897-71a8-4b61-a9be-4d963e1384ac\") " pod="openstack-operators/openstack-operator-controller-operator-7bfcb77fcc-bx765" Dec 04 04:10:09 crc kubenswrapper[4806]: I1204 04:10:09.957689 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h5d4\" (UniqueName: \"kubernetes.io/projected/4d229897-71a8-4b61-a9be-4d963e1384ac-kube-api-access-8h5d4\") pod \"openstack-operator-controller-operator-7bfcb77fcc-bx765\" (UID: \"4d229897-71a8-4b61-a9be-4d963e1384ac\") " pod="openstack-operators/openstack-operator-controller-operator-7bfcb77fcc-bx765" Dec 04 04:10:09 crc kubenswrapper[4806]: I1204 04:10:09.989800 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7bfcb77fcc-bx765" Dec 04 04:10:10 crc kubenswrapper[4806]: I1204 04:10:10.241532 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7bfcb77fcc-bx765"] Dec 04 04:10:10 crc kubenswrapper[4806]: I1204 04:10:10.335773 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7bfcb77fcc-bx765" event={"ID":"4d229897-71a8-4b61-a9be-4d963e1384ac","Type":"ContainerStarted","Data":"afdb4b1e7c8aea242b448b81875a8023896e035d8c6ef863c722a7d31da6be56"} Dec 04 04:10:16 crc kubenswrapper[4806]: I1204 04:10:16.471582 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7bfcb77fcc-bx765" event={"ID":"4d229897-71a8-4b61-a9be-4d963e1384ac","Type":"ContainerStarted","Data":"920507406fecbe94cd5f30c28c259bf99793fc6489738872a1f52cb8ddfa34c3"} Dec 04 04:10:16 crc kubenswrapper[4806]: I1204 04:10:16.472180 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7bfcb77fcc-bx765" Dec 04 04:10:16 crc kubenswrapper[4806]: I1204 04:10:16.510861 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7bfcb77fcc-bx765" podStartSLOduration=2.23126514 podStartE2EDuration="7.510841025s" podCreationTimestamp="2025-12-04 04:10:09 +0000 UTC" firstStartedPulling="2025-12-04 04:10:10.248775703 +0000 UTC m=+925.107288631" lastFinishedPulling="2025-12-04 04:10:15.528351588 +0000 UTC m=+930.386864516" observedRunningTime="2025-12-04 04:10:16.506760176 +0000 UTC m=+931.365273104" watchObservedRunningTime="2025-12-04 04:10:16.510841025 +0000 UTC m=+931.369353953" Dec 04 04:10:19 crc kubenswrapper[4806]: I1204 04:10:19.591137 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-22j88"] Dec 04 04:10:19 crc kubenswrapper[4806]: I1204 04:10:19.592774 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:19 crc kubenswrapper[4806]: I1204 04:10:19.597161 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aff17e3a-3656-442c-bdd7-c029c178edec-utilities\") pod \"redhat-marketplace-22j88\" (UID: \"aff17e3a-3656-442c-bdd7-c029c178edec\") " pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:19 crc kubenswrapper[4806]: I1204 04:10:19.597241 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6slx\" (UniqueName: \"kubernetes.io/projected/aff17e3a-3656-442c-bdd7-c029c178edec-kube-api-access-f6slx\") pod \"redhat-marketplace-22j88\" (UID: \"aff17e3a-3656-442c-bdd7-c029c178edec\") " pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:19 crc kubenswrapper[4806]: I1204 04:10:19.597279 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aff17e3a-3656-442c-bdd7-c029c178edec-catalog-content\") pod \"redhat-marketplace-22j88\" (UID: \"aff17e3a-3656-442c-bdd7-c029c178edec\") " pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:19 crc kubenswrapper[4806]: I1204 04:10:19.616873 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-22j88"] Dec 04 04:10:19 crc kubenswrapper[4806]: I1204 04:10:19.698130 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aff17e3a-3656-442c-bdd7-c029c178edec-utilities\") pod \"redhat-marketplace-22j88\" (UID: \"aff17e3a-3656-442c-bdd7-c029c178edec\") " pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:19 crc kubenswrapper[4806]: I1204 04:10:19.698398 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6slx\" (UniqueName: \"kubernetes.io/projected/aff17e3a-3656-442c-bdd7-c029c178edec-kube-api-access-f6slx\") pod \"redhat-marketplace-22j88\" (UID: \"aff17e3a-3656-442c-bdd7-c029c178edec\") " pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:19 crc kubenswrapper[4806]: I1204 04:10:19.698527 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aff17e3a-3656-442c-bdd7-c029c178edec-catalog-content\") pod \"redhat-marketplace-22j88\" (UID: \"aff17e3a-3656-442c-bdd7-c029c178edec\") " pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:19 crc kubenswrapper[4806]: I1204 04:10:19.698661 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aff17e3a-3656-442c-bdd7-c029c178edec-utilities\") pod \"redhat-marketplace-22j88\" (UID: \"aff17e3a-3656-442c-bdd7-c029c178edec\") " pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:19 crc kubenswrapper[4806]: I1204 04:10:19.698981 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aff17e3a-3656-442c-bdd7-c029c178edec-catalog-content\") pod \"redhat-marketplace-22j88\" (UID: \"aff17e3a-3656-442c-bdd7-c029c178edec\") " pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:19 crc kubenswrapper[4806]: I1204 04:10:19.718317 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6slx\" (UniqueName: \"kubernetes.io/projected/aff17e3a-3656-442c-bdd7-c029c178edec-kube-api-access-f6slx\") pod \"redhat-marketplace-22j88\" (UID: \"aff17e3a-3656-442c-bdd7-c029c178edec\") " pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:19 crc kubenswrapper[4806]: I1204 04:10:19.908383 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:20 crc kubenswrapper[4806]: I1204 04:10:20.182222 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-22j88"] Dec 04 04:10:20 crc kubenswrapper[4806]: I1204 04:10:20.501628 4806 generic.go:334] "Generic (PLEG): container finished" podID="aff17e3a-3656-442c-bdd7-c029c178edec" containerID="02eaa7c81b812e603f7cfebbf72942ca7e046d6b1532aa99d8fb9b2e280fc0af" exitCode=0 Dec 04 04:10:20 crc kubenswrapper[4806]: I1204 04:10:20.501683 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22j88" event={"ID":"aff17e3a-3656-442c-bdd7-c029c178edec","Type":"ContainerDied","Data":"02eaa7c81b812e603f7cfebbf72942ca7e046d6b1532aa99d8fb9b2e280fc0af"} Dec 04 04:10:20 crc kubenswrapper[4806]: I1204 04:10:20.501710 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22j88" event={"ID":"aff17e3a-3656-442c-bdd7-c029c178edec","Type":"ContainerStarted","Data":"b50228895ad978f40352e7ad8b01a6502190a0f20aed49bac6082077e2eb8df6"} Dec 04 04:10:21 crc kubenswrapper[4806]: I1204 04:10:21.508052 4806 generic.go:334] "Generic (PLEG): container finished" podID="aff17e3a-3656-442c-bdd7-c029c178edec" containerID="0dc87cd3b094c50007e0baab9a4b5497b29e6fab4aee9a638c369235158f8f6a" exitCode=0 Dec 04 04:10:21 crc kubenswrapper[4806]: I1204 04:10:21.508386 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22j88" event={"ID":"aff17e3a-3656-442c-bdd7-c029c178edec","Type":"ContainerDied","Data":"0dc87cd3b094c50007e0baab9a4b5497b29e6fab4aee9a638c369235158f8f6a"} Dec 04 04:10:22 crc kubenswrapper[4806]: I1204 04:10:22.515770 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22j88" event={"ID":"aff17e3a-3656-442c-bdd7-c029c178edec","Type":"ContainerStarted","Data":"60e906fa1793568ed6ddfe8049596b6c7d82abcb14c4f6c563b28a5e4f15ce40"} Dec 04 04:10:22 crc kubenswrapper[4806]: I1204 04:10:22.535114 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-22j88" podStartSLOduration=2.1737988 podStartE2EDuration="3.535095031s" podCreationTimestamp="2025-12-04 04:10:19 +0000 UTC" firstStartedPulling="2025-12-04 04:10:20.503298212 +0000 UTC m=+935.361811140" lastFinishedPulling="2025-12-04 04:10:21.864594443 +0000 UTC m=+936.723107371" observedRunningTime="2025-12-04 04:10:22.531974093 +0000 UTC m=+937.390487041" watchObservedRunningTime="2025-12-04 04:10:22.535095031 +0000 UTC m=+937.393607959" Dec 04 04:10:27 crc kubenswrapper[4806]: I1204 04:10:27.047429 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:10:27 crc kubenswrapper[4806]: I1204 04:10:27.047787 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:10:27 crc kubenswrapper[4806]: I1204 04:10:27.047835 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 04:10:27 crc kubenswrapper[4806]: I1204 04:10:27.048478 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4ce73f616eb3f88f0847c6436643d27fc2337de01de1b065778733e8ceef4522"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 04:10:27 crc kubenswrapper[4806]: I1204 04:10:27.048532 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://4ce73f616eb3f88f0847c6436643d27fc2337de01de1b065778733e8ceef4522" gracePeriod=600 Dec 04 04:10:27 crc kubenswrapper[4806]: E1204 04:10:27.517316 4806 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0f9623b_44ce_43c2_b990_a75db2d17ff8.slice/crio-conmon-4ce73f616eb3f88f0847c6436643d27fc2337de01de1b065778733e8ceef4522.scope\": RecentStats: unable to find data in memory cache]" Dec 04 04:10:27 crc kubenswrapper[4806]: I1204 04:10:27.546618 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="4ce73f616eb3f88f0847c6436643d27fc2337de01de1b065778733e8ceef4522" exitCode=0 Dec 04 04:10:27 crc kubenswrapper[4806]: I1204 04:10:27.546661 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"4ce73f616eb3f88f0847c6436643d27fc2337de01de1b065778733e8ceef4522"} Dec 04 04:10:27 crc kubenswrapper[4806]: I1204 04:10:27.546694 4806 scope.go:117] "RemoveContainer" containerID="f0bc0378fad2cb0fb66c2e5ff86d10f9a724302404985dba5e351fcfd4397778" Dec 04 04:10:28 crc kubenswrapper[4806]: I1204 04:10:28.556308 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"ff51617027055d5b0c7e244b302f5ed7ba3e3cbe117dc54d0ea82947af1a87f2"} Dec 04 04:10:29 crc kubenswrapper[4806]: I1204 04:10:29.908513 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:29 crc kubenswrapper[4806]: I1204 04:10:29.908894 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:29 crc kubenswrapper[4806]: I1204 04:10:29.952492 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:29 crc kubenswrapper[4806]: I1204 04:10:29.993111 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7bfcb77fcc-bx765" Dec 04 04:10:30 crc kubenswrapper[4806]: I1204 04:10:30.608576 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:30 crc kubenswrapper[4806]: I1204 04:10:30.649982 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-22j88"] Dec 04 04:10:32 crc kubenswrapper[4806]: I1204 04:10:32.575700 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-22j88" podUID="aff17e3a-3656-442c-bdd7-c029c178edec" containerName="registry-server" containerID="cri-o://60e906fa1793568ed6ddfe8049596b6c7d82abcb14c4f6c563b28a5e4f15ce40" gracePeriod=2 Dec 04 04:10:32 crc kubenswrapper[4806]: I1204 04:10:32.973990 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.104296 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6slx\" (UniqueName: \"kubernetes.io/projected/aff17e3a-3656-442c-bdd7-c029c178edec-kube-api-access-f6slx\") pod \"aff17e3a-3656-442c-bdd7-c029c178edec\" (UID: \"aff17e3a-3656-442c-bdd7-c029c178edec\") " Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.104451 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aff17e3a-3656-442c-bdd7-c029c178edec-catalog-content\") pod \"aff17e3a-3656-442c-bdd7-c029c178edec\" (UID: \"aff17e3a-3656-442c-bdd7-c029c178edec\") " Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.104551 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aff17e3a-3656-442c-bdd7-c029c178edec-utilities\") pod \"aff17e3a-3656-442c-bdd7-c029c178edec\" (UID: \"aff17e3a-3656-442c-bdd7-c029c178edec\") " Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.105349 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aff17e3a-3656-442c-bdd7-c029c178edec-utilities" (OuterVolumeSpecName: "utilities") pod "aff17e3a-3656-442c-bdd7-c029c178edec" (UID: "aff17e3a-3656-442c-bdd7-c029c178edec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.122149 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aff17e3a-3656-442c-bdd7-c029c178edec-kube-api-access-f6slx" (OuterVolumeSpecName: "kube-api-access-f6slx") pod "aff17e3a-3656-442c-bdd7-c029c178edec" (UID: "aff17e3a-3656-442c-bdd7-c029c178edec"). InnerVolumeSpecName "kube-api-access-f6slx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.124862 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aff17e3a-3656-442c-bdd7-c029c178edec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aff17e3a-3656-442c-bdd7-c029c178edec" (UID: "aff17e3a-3656-442c-bdd7-c029c178edec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.206172 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6slx\" (UniqueName: \"kubernetes.io/projected/aff17e3a-3656-442c-bdd7-c029c178edec-kube-api-access-f6slx\") on node \"crc\" DevicePath \"\"" Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.206224 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aff17e3a-3656-442c-bdd7-c029c178edec-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.206239 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aff17e3a-3656-442c-bdd7-c029c178edec-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.582742 4806 generic.go:334] "Generic (PLEG): container finished" podID="aff17e3a-3656-442c-bdd7-c029c178edec" containerID="60e906fa1793568ed6ddfe8049596b6c7d82abcb14c4f6c563b28a5e4f15ce40" exitCode=0 Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.582818 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-22j88" Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.582817 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22j88" event={"ID":"aff17e3a-3656-442c-bdd7-c029c178edec","Type":"ContainerDied","Data":"60e906fa1793568ed6ddfe8049596b6c7d82abcb14c4f6c563b28a5e4f15ce40"} Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.583244 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-22j88" event={"ID":"aff17e3a-3656-442c-bdd7-c029c178edec","Type":"ContainerDied","Data":"b50228895ad978f40352e7ad8b01a6502190a0f20aed49bac6082077e2eb8df6"} Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.583268 4806 scope.go:117] "RemoveContainer" containerID="60e906fa1793568ed6ddfe8049596b6c7d82abcb14c4f6c563b28a5e4f15ce40" Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.607580 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-22j88"] Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.612322 4806 scope.go:117] "RemoveContainer" containerID="0dc87cd3b094c50007e0baab9a4b5497b29e6fab4aee9a638c369235158f8f6a" Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.613736 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-22j88"] Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.628099 4806 scope.go:117] "RemoveContainer" containerID="02eaa7c81b812e603f7cfebbf72942ca7e046d6b1532aa99d8fb9b2e280fc0af" Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.649909 4806 scope.go:117] "RemoveContainer" containerID="60e906fa1793568ed6ddfe8049596b6c7d82abcb14c4f6c563b28a5e4f15ce40" Dec 04 04:10:33 crc kubenswrapper[4806]: E1204 04:10:33.650489 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60e906fa1793568ed6ddfe8049596b6c7d82abcb14c4f6c563b28a5e4f15ce40\": container with ID starting with 60e906fa1793568ed6ddfe8049596b6c7d82abcb14c4f6c563b28a5e4f15ce40 not found: ID does not exist" containerID="60e906fa1793568ed6ddfe8049596b6c7d82abcb14c4f6c563b28a5e4f15ce40" Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.650668 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60e906fa1793568ed6ddfe8049596b6c7d82abcb14c4f6c563b28a5e4f15ce40"} err="failed to get container status \"60e906fa1793568ed6ddfe8049596b6c7d82abcb14c4f6c563b28a5e4f15ce40\": rpc error: code = NotFound desc = could not find container \"60e906fa1793568ed6ddfe8049596b6c7d82abcb14c4f6c563b28a5e4f15ce40\": container with ID starting with 60e906fa1793568ed6ddfe8049596b6c7d82abcb14c4f6c563b28a5e4f15ce40 not found: ID does not exist" Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.650778 4806 scope.go:117] "RemoveContainer" containerID="0dc87cd3b094c50007e0baab9a4b5497b29e6fab4aee9a638c369235158f8f6a" Dec 04 04:10:33 crc kubenswrapper[4806]: E1204 04:10:33.651230 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dc87cd3b094c50007e0baab9a4b5497b29e6fab4aee9a638c369235158f8f6a\": container with ID starting with 0dc87cd3b094c50007e0baab9a4b5497b29e6fab4aee9a638c369235158f8f6a not found: ID does not exist" containerID="0dc87cd3b094c50007e0baab9a4b5497b29e6fab4aee9a638c369235158f8f6a" Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.651263 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dc87cd3b094c50007e0baab9a4b5497b29e6fab4aee9a638c369235158f8f6a"} err="failed to get container status \"0dc87cd3b094c50007e0baab9a4b5497b29e6fab4aee9a638c369235158f8f6a\": rpc error: code = NotFound desc = could not find container \"0dc87cd3b094c50007e0baab9a4b5497b29e6fab4aee9a638c369235158f8f6a\": container with ID starting with 0dc87cd3b094c50007e0baab9a4b5497b29e6fab4aee9a638c369235158f8f6a not found: ID does not exist" Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.651283 4806 scope.go:117] "RemoveContainer" containerID="02eaa7c81b812e603f7cfebbf72942ca7e046d6b1532aa99d8fb9b2e280fc0af" Dec 04 04:10:33 crc kubenswrapper[4806]: E1204 04:10:33.651611 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02eaa7c81b812e603f7cfebbf72942ca7e046d6b1532aa99d8fb9b2e280fc0af\": container with ID starting with 02eaa7c81b812e603f7cfebbf72942ca7e046d6b1532aa99d8fb9b2e280fc0af not found: ID does not exist" containerID="02eaa7c81b812e603f7cfebbf72942ca7e046d6b1532aa99d8fb9b2e280fc0af" Dec 04 04:10:33 crc kubenswrapper[4806]: I1204 04:10:33.651695 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02eaa7c81b812e603f7cfebbf72942ca7e046d6b1532aa99d8fb9b2e280fc0af"} err="failed to get container status \"02eaa7c81b812e603f7cfebbf72942ca7e046d6b1532aa99d8fb9b2e280fc0af\": rpc error: code = NotFound desc = could not find container \"02eaa7c81b812e603f7cfebbf72942ca7e046d6b1532aa99d8fb9b2e280fc0af\": container with ID starting with 02eaa7c81b812e603f7cfebbf72942ca7e046d6b1532aa99d8fb9b2e280fc0af not found: ID does not exist" Dec 04 04:10:35 crc kubenswrapper[4806]: I1204 04:10:35.437700 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aff17e3a-3656-442c-bdd7-c029c178edec" path="/var/lib/kubelet/pods/aff17e3a-3656-442c-bdd7-c029c178edec/volumes" Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.261879 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tpq2v"] Dec 04 04:10:44 crc kubenswrapper[4806]: E1204 04:10:44.262559 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aff17e3a-3656-442c-bdd7-c029c178edec" containerName="extract-content" Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.262571 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="aff17e3a-3656-442c-bdd7-c029c178edec" containerName="extract-content" Dec 04 04:10:44 crc kubenswrapper[4806]: E1204 04:10:44.262579 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aff17e3a-3656-442c-bdd7-c029c178edec" containerName="extract-utilities" Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.262586 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="aff17e3a-3656-442c-bdd7-c029c178edec" containerName="extract-utilities" Dec 04 04:10:44 crc kubenswrapper[4806]: E1204 04:10:44.262604 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aff17e3a-3656-442c-bdd7-c029c178edec" containerName="registry-server" Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.262610 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="aff17e3a-3656-442c-bdd7-c029c178edec" containerName="registry-server" Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.262713 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="aff17e3a-3656-442c-bdd7-c029c178edec" containerName="registry-server" Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.263485 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.284719 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tpq2v"] Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.432415 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-utilities\") pod \"community-operators-tpq2v\" (UID: \"8cb9a7f5-2b1d-436d-9c56-b0898279ba66\") " pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.432542 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lf6r\" (UniqueName: \"kubernetes.io/projected/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-kube-api-access-4lf6r\") pod \"community-operators-tpq2v\" (UID: \"8cb9a7f5-2b1d-436d-9c56-b0898279ba66\") " pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.432573 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-catalog-content\") pod \"community-operators-tpq2v\" (UID: \"8cb9a7f5-2b1d-436d-9c56-b0898279ba66\") " pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.533159 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lf6r\" (UniqueName: \"kubernetes.io/projected/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-kube-api-access-4lf6r\") pod \"community-operators-tpq2v\" (UID: \"8cb9a7f5-2b1d-436d-9c56-b0898279ba66\") " pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.533210 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-catalog-content\") pod \"community-operators-tpq2v\" (UID: \"8cb9a7f5-2b1d-436d-9c56-b0898279ba66\") " pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.533240 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-utilities\") pod \"community-operators-tpq2v\" (UID: \"8cb9a7f5-2b1d-436d-9c56-b0898279ba66\") " pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.533738 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-utilities\") pod \"community-operators-tpq2v\" (UID: \"8cb9a7f5-2b1d-436d-9c56-b0898279ba66\") " pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.533782 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-catalog-content\") pod \"community-operators-tpq2v\" (UID: \"8cb9a7f5-2b1d-436d-9c56-b0898279ba66\") " pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.555104 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lf6r\" (UniqueName: \"kubernetes.io/projected/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-kube-api-access-4lf6r\") pod \"community-operators-tpq2v\" (UID: \"8cb9a7f5-2b1d-436d-9c56-b0898279ba66\") " pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:44 crc kubenswrapper[4806]: I1204 04:10:44.579205 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:45 crc kubenswrapper[4806]: I1204 04:10:45.154015 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tpq2v"] Dec 04 04:10:45 crc kubenswrapper[4806]: I1204 04:10:45.675786 4806 generic.go:334] "Generic (PLEG): container finished" podID="8cb9a7f5-2b1d-436d-9c56-b0898279ba66" containerID="79dbef329e4746d2879807f0eb67ae2fc37b6b7dc74f4b102ea6c7e58c6507e9" exitCode=0 Dec 04 04:10:45 crc kubenswrapper[4806]: I1204 04:10:45.675861 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpq2v" event={"ID":"8cb9a7f5-2b1d-436d-9c56-b0898279ba66","Type":"ContainerDied","Data":"79dbef329e4746d2879807f0eb67ae2fc37b6b7dc74f4b102ea6c7e58c6507e9"} Dec 04 04:10:45 crc kubenswrapper[4806]: I1204 04:10:45.676249 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpq2v" event={"ID":"8cb9a7f5-2b1d-436d-9c56-b0898279ba66","Type":"ContainerStarted","Data":"82cef718e3ba375835bc15f8250fc69c841e8b299151a39e391dd1b72f925d4e"} Dec 04 04:10:47 crc kubenswrapper[4806]: I1204 04:10:47.690752 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpq2v" event={"ID":"8cb9a7f5-2b1d-436d-9c56-b0898279ba66","Type":"ContainerStarted","Data":"b1cbad48e2334478d378dc4810cfd11f38ed6364195e9f9c5168e23ecf12708e"} Dec 04 04:10:49 crc kubenswrapper[4806]: I1204 04:10:49.703694 4806 generic.go:334] "Generic (PLEG): container finished" podID="8cb9a7f5-2b1d-436d-9c56-b0898279ba66" containerID="b1cbad48e2334478d378dc4810cfd11f38ed6364195e9f9c5168e23ecf12708e" exitCode=0 Dec 04 04:10:49 crc kubenswrapper[4806]: I1204 04:10:49.703781 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpq2v" event={"ID":"8cb9a7f5-2b1d-436d-9c56-b0898279ba66","Type":"ContainerDied","Data":"b1cbad48e2334478d378dc4810cfd11f38ed6364195e9f9c5168e23ecf12708e"} Dec 04 04:10:50 crc kubenswrapper[4806]: I1204 04:10:50.716091 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpq2v" event={"ID":"8cb9a7f5-2b1d-436d-9c56-b0898279ba66","Type":"ContainerStarted","Data":"f925a91b8b14fa816810bc818f02263714cea10c2802c9de63748c696a514385"} Dec 04 04:10:51 crc kubenswrapper[4806]: I1204 04:10:51.033651 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tpq2v" podStartSLOduration=2.563950319 podStartE2EDuration="7.033631167s" podCreationTimestamp="2025-12-04 04:10:44 +0000 UTC" firstStartedPulling="2025-12-04 04:10:45.677687569 +0000 UTC m=+960.536200497" lastFinishedPulling="2025-12-04 04:10:50.147368417 +0000 UTC m=+965.005881345" observedRunningTime="2025-12-04 04:10:51.029643483 +0000 UTC m=+965.888156401" watchObservedRunningTime="2025-12-04 04:10:51.033631167 +0000 UTC m=+965.892144095" Dec 04 04:10:54 crc kubenswrapper[4806]: I1204 04:10:54.581443 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:54 crc kubenswrapper[4806]: I1204 04:10:54.584178 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:54 crc kubenswrapper[4806]: I1204 04:10:54.642349 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:55 crc kubenswrapper[4806]: I1204 04:10:55.809255 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:55 crc kubenswrapper[4806]: I1204 04:10:55.859436 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tpq2v"] Dec 04 04:10:57 crc kubenswrapper[4806]: I1204 04:10:57.754201 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tpq2v" podUID="8cb9a7f5-2b1d-436d-9c56-b0898279ba66" containerName="registry-server" containerID="cri-o://f925a91b8b14fa816810bc818f02263714cea10c2802c9de63748c696a514385" gracePeriod=2 Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.413092 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-g2dfw"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.416819 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g2dfw" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.423111 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-g2dfw"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.439528 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-jvc9q" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.442073 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-25nxx"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.443223 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-25nxx" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.448168 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-m27xk" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.492335 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-25nxx"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.496666 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-mxmf2"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.497697 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mxmf2" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.499487 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-4dprl" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.501749 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-xd6pj"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.502724 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-xd6pj" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.504207 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-nnrnm" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.518063 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-mxmf2"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.533089 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzrwg\" (UniqueName: \"kubernetes.io/projected/b2bba37e-d570-432f-9520-9b5de2005e5c-kube-api-access-jzrwg\") pod \"barbican-operator-controller-manager-7d9dfd778-g2dfw\" (UID: \"b2bba37e-d570-432f-9520-9b5de2005e5c\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g2dfw" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.553532 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vhxjt"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.554544 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vhxjt" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.559098 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-4qnl9" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.563990 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vtcg5"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.565016 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vtcg5" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.569870 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-qj6cz" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.594852 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vtcg5"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.625014 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-g6svq"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.626241 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.636382 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-n6tk8" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.636566 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.637469 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzlmc\" (UniqueName: \"kubernetes.io/projected/91ab6e01-027e-4e7a-a250-374b4e8873fc-kube-api-access-wzlmc\") pod \"designate-operator-controller-manager-78b4bc895b-mxmf2\" (UID: \"91ab6e01-027e-4e7a-a250-374b4e8873fc\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mxmf2" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.637516 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cddnt\" (UniqueName: \"kubernetes.io/projected/7213e0bc-21af-4df7-929d-f915e343f184-kube-api-access-cddnt\") pod \"glance-operator-controller-manager-77987cd8cd-xd6pj\" (UID: \"7213e0bc-21af-4df7-929d-f915e343f184\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-xd6pj" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.637546 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzrwg\" (UniqueName: \"kubernetes.io/projected/b2bba37e-d570-432f-9520-9b5de2005e5c-kube-api-access-jzrwg\") pod \"barbican-operator-controller-manager-7d9dfd778-g2dfw\" (UID: \"b2bba37e-d570-432f-9520-9b5de2005e5c\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g2dfw" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.637584 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh48x\" (UniqueName: \"kubernetes.io/projected/a022d699-4a2b-49e2-8d8f-787a3d888df8-kube-api-access-gh48x\") pod \"cinder-operator-controller-manager-859b6ccc6-25nxx\" (UID: \"a022d699-4a2b-49e2-8d8f-787a3d888df8\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-25nxx" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.666698 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzrwg\" (UniqueName: \"kubernetes.io/projected/b2bba37e-d570-432f-9520-9b5de2005e5c-kube-api-access-jzrwg\") pod \"barbican-operator-controller-manager-7d9dfd778-g2dfw\" (UID: \"b2bba37e-d570-432f-9520-9b5de2005e5c\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g2dfw" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.671385 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-xd6pj"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.695274 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-g6svq"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.724973 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-f6k5f"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.739430 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh48x\" (UniqueName: \"kubernetes.io/projected/a022d699-4a2b-49e2-8d8f-787a3d888df8-kube-api-access-gh48x\") pod \"cinder-operator-controller-manager-859b6ccc6-25nxx\" (UID: \"a022d699-4a2b-49e2-8d8f-787a3d888df8\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-25nxx" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.739790 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert\") pod \"infra-operator-controller-manager-57548d458d-g6svq\" (UID: \"04d2c949-c17a-4299-a9be-b56a7c34866c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.746194 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lftm\" (UniqueName: \"kubernetes.io/projected/04d2c949-c17a-4299-a9be-b56a7c34866c-kube-api-access-8lftm\") pod \"infra-operator-controller-manager-57548d458d-g6svq\" (UID: \"04d2c949-c17a-4299-a9be-b56a7c34866c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.746357 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzlmc\" (UniqueName: \"kubernetes.io/projected/91ab6e01-027e-4e7a-a250-374b4e8873fc-kube-api-access-wzlmc\") pod \"designate-operator-controller-manager-78b4bc895b-mxmf2\" (UID: \"91ab6e01-027e-4e7a-a250-374b4e8873fc\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mxmf2" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.746542 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4dzb\" (UniqueName: \"kubernetes.io/projected/b496f26d-2493-416f-a0f8-5b33e41dbda0-kube-api-access-s4dzb\") pod \"heat-operator-controller-manager-5f64f6f8bb-vhxjt\" (UID: \"b496f26d-2493-416f-a0f8-5b33e41dbda0\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vhxjt" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.746654 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cddnt\" (UniqueName: \"kubernetes.io/projected/7213e0bc-21af-4df7-929d-f915e343f184-kube-api-access-cddnt\") pod \"glance-operator-controller-manager-77987cd8cd-xd6pj\" (UID: \"7213e0bc-21af-4df7-929d-f915e343f184\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-xd6pj" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.746772 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqkgr\" (UniqueName: \"kubernetes.io/projected/084c526b-94d8-43ae-9ffb-33aa215de8c2-kube-api-access-dqkgr\") pod \"horizon-operator-controller-manager-68c6d99b8f-vtcg5\" (UID: \"084c526b-94d8-43ae-9ffb-33aa215de8c2\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vtcg5" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.743838 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-f6k5f" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.748484 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g2dfw" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.751551 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-8dvqz" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.754212 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-f6k5f"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.785971 4806 generic.go:334] "Generic (PLEG): container finished" podID="8cb9a7f5-2b1d-436d-9c56-b0898279ba66" containerID="f925a91b8b14fa816810bc818f02263714cea10c2802c9de63748c696a514385" exitCode=0 Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.786895 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpq2v" event={"ID":"8cb9a7f5-2b1d-436d-9c56-b0898279ba66","Type":"ContainerDied","Data":"f925a91b8b14fa816810bc818f02263714cea10c2802c9de63748c696a514385"} Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.836277 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzlmc\" (UniqueName: \"kubernetes.io/projected/91ab6e01-027e-4e7a-a250-374b4e8873fc-kube-api-access-wzlmc\") pod \"designate-operator-controller-manager-78b4bc895b-mxmf2\" (UID: \"91ab6e01-027e-4e7a-a250-374b4e8873fc\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mxmf2" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.792578 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cddnt\" (UniqueName: \"kubernetes.io/projected/7213e0bc-21af-4df7-929d-f915e343f184-kube-api-access-cddnt\") pod \"glance-operator-controller-manager-77987cd8cd-xd6pj\" (UID: \"7213e0bc-21af-4df7-929d-f915e343f184\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-xd6pj" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.843186 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-xd6pj" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.850820 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4dzb\" (UniqueName: \"kubernetes.io/projected/b496f26d-2493-416f-a0f8-5b33e41dbda0-kube-api-access-s4dzb\") pod \"heat-operator-controller-manager-5f64f6f8bb-vhxjt\" (UID: \"b496f26d-2493-416f-a0f8-5b33e41dbda0\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vhxjt" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.850882 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqkgr\" (UniqueName: \"kubernetes.io/projected/084c526b-94d8-43ae-9ffb-33aa215de8c2-kube-api-access-dqkgr\") pod \"horizon-operator-controller-manager-68c6d99b8f-vtcg5\" (UID: \"084c526b-94d8-43ae-9ffb-33aa215de8c2\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vtcg5" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.850959 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert\") pod \"infra-operator-controller-manager-57548d458d-g6svq\" (UID: \"04d2c949-c17a-4299-a9be-b56a7c34866c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.851022 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lftm\" (UniqueName: \"kubernetes.io/projected/04d2c949-c17a-4299-a9be-b56a7c34866c-kube-api-access-8lftm\") pod \"infra-operator-controller-manager-57548d458d-g6svq\" (UID: \"04d2c949-c17a-4299-a9be-b56a7c34866c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.851076 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttnjh\" (UniqueName: \"kubernetes.io/projected/5f55ddcf-ca5d-4249-9384-e541ae7d4536-kube-api-access-ttnjh\") pod \"ironic-operator-controller-manager-6c548fd776-f6k5f\" (UID: \"5f55ddcf-ca5d-4249-9384-e541ae7d4536\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-f6k5f" Dec 04 04:10:58 crc kubenswrapper[4806]: E1204 04:10:58.851521 4806 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 04:10:58 crc kubenswrapper[4806]: E1204 04:10:58.851562 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert podName:04d2c949-c17a-4299-a9be-b56a7c34866c nodeName:}" failed. No retries permitted until 2025-12-04 04:10:59.351547707 +0000 UTC m=+974.210060635 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert") pod "infra-operator-controller-manager-57548d458d-g6svq" (UID: "04d2c949-c17a-4299-a9be-b56a7c34866c") : secret "infra-operator-webhook-server-cert" not found Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.861029 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh48x\" (UniqueName: \"kubernetes.io/projected/a022d699-4a2b-49e2-8d8f-787a3d888df8-kube-api-access-gh48x\") pod \"cinder-operator-controller-manager-859b6ccc6-25nxx\" (UID: \"a022d699-4a2b-49e2-8d8f-787a3d888df8\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-25nxx" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.867413 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-nt5vq"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.868548 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nt5vq" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.879577 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-hnbpd" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.911746 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lftm\" (UniqueName: \"kubernetes.io/projected/04d2c949-c17a-4299-a9be-b56a7c34866c-kube-api-access-8lftm\") pod \"infra-operator-controller-manager-57548d458d-g6svq\" (UID: \"04d2c949-c17a-4299-a9be-b56a7c34866c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.911915 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vhxjt"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.920465 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqkgr\" (UniqueName: \"kubernetes.io/projected/084c526b-94d8-43ae-9ffb-33aa215de8c2-kube-api-access-dqkgr\") pod \"horizon-operator-controller-manager-68c6d99b8f-vtcg5\" (UID: \"084c526b-94d8-43ae-9ffb-33aa215de8c2\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vtcg5" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.926630 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4dzb\" (UniqueName: \"kubernetes.io/projected/b496f26d-2493-416f-a0f8-5b33e41dbda0-kube-api-access-s4dzb\") pod \"heat-operator-controller-manager-5f64f6f8bb-vhxjt\" (UID: \"b496f26d-2493-416f-a0f8-5b33e41dbda0\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vhxjt" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.952336 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttnjh\" (UniqueName: \"kubernetes.io/projected/5f55ddcf-ca5d-4249-9384-e541ae7d4536-kube-api-access-ttnjh\") pod \"ironic-operator-controller-manager-6c548fd776-f6k5f\" (UID: \"5f55ddcf-ca5d-4249-9384-e541ae7d4536\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-f6k5f" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.962002 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-2spbg"] Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.963506 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2spbg" Dec 04 04:10:58 crc kubenswrapper[4806]: I1204 04:10:58.984980 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-qhs26" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.017821 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-nt5vq"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.034967 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-45jzr"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.036143 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-45jzr" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.050312 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttnjh\" (UniqueName: \"kubernetes.io/projected/5f55ddcf-ca5d-4249-9384-e541ae7d4536-kube-api-access-ttnjh\") pod \"ironic-operator-controller-manager-6c548fd776-f6k5f\" (UID: \"5f55ddcf-ca5d-4249-9384-e541ae7d4536\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-f6k5f" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.055376 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-2spbg"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.056040 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lx4r\" (UniqueName: \"kubernetes.io/projected/669dd56a-69b0-4315-aacc-c8f085f0b76a-kube-api-access-7lx4r\") pod \"keystone-operator-controller-manager-7765d96ddf-nt5vq\" (UID: \"669dd56a-69b0-4315-aacc-c8f085f0b76a\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nt5vq" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.061845 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-25nxx" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.069325 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-nbrlg" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.090621 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-45jzr"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.098330 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2b2qk"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.099401 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2b2qk" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.111221 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2b2qk"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.126263 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-pk7ws"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.127103 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-pk7ws"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.127181 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-pk7ws" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.134515 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-kcmvz"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.135504 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kcmvz" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.136193 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mxmf2" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.138511 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-kcmvz"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.143117 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-sktq6"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.144125 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sktq6" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.152224 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-bvpcr" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.152345 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-wv4jq" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.152478 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.152561 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-rb9f4" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.152875 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-8hv9q" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.153483 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.154057 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-f6k5f" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.160256 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-sktq6"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.161597 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lx4r\" (UniqueName: \"kubernetes.io/projected/669dd56a-69b0-4315-aacc-c8f085f0b76a-kube-api-access-7lx4r\") pod \"keystone-operator-controller-manager-7765d96ddf-nt5vq\" (UID: \"669dd56a-69b0-4315-aacc-c8f085f0b76a\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nt5vq" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.161644 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn4rx\" (UniqueName: \"kubernetes.io/projected/95ae1fa1-3ac7-4040-86da-150daca97f6b-kube-api-access-gn4rx\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-2b2qk\" (UID: \"95ae1fa1-3ac7-4040-86da-150daca97f6b\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2b2qk" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.161678 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t94cj\" (UniqueName: \"kubernetes.io/projected/310d58eb-6345-4cba-8a01-373c65c7bc40-kube-api-access-t94cj\") pod \"mariadb-operator-controller-manager-56bbcc9d85-45jzr\" (UID: \"310d58eb-6345-4cba-8a01-373c65c7bc40\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-45jzr" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.161701 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whrgb\" (UniqueName: \"kubernetes.io/projected/5f479a93-bb3a-45b8-aa74-60e9f0a7aebd-kube-api-access-whrgb\") pod \"nova-operator-controller-manager-697bc559fc-pk7ws\" (UID: \"5f479a93-bb3a-45b8-aa74-60e9f0a7aebd\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-pk7ws" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.161721 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-778dw\" (UniqueName: \"kubernetes.io/projected/1cfe7f94-e830-4145-a3e7-8a3f694adda1-kube-api-access-778dw\") pod \"octavia-operator-controller-manager-998648c74-kcmvz\" (UID: \"1cfe7f94-e830-4145-a3e7-8a3f694adda1\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-kcmvz" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.161749 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j29nf\" (UniqueName: \"kubernetes.io/projected/2a7907e6-c36f-4534-9460-80c0091af286-kube-api-access-j29nf\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7\" (UID: \"2a7907e6-c36f-4534-9460-80c0091af286\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.161774 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jffl2\" (UniqueName: \"kubernetes.io/projected/66428f64-e34e-4b37-b911-13e7142f1c18-kube-api-access-jffl2\") pod \"manila-operator-controller-manager-7c79b5df47-2spbg\" (UID: \"66428f64-e34e-4b37-b911-13e7142f1c18\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2spbg" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.161792 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7\" (UID: \"2a7907e6-c36f-4534-9460-80c0091af286\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.161816 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj5q9\" (UniqueName: \"kubernetes.io/projected/e06aa87c-9f7d-44f7-b96d-27141d2fff93-kube-api-access-lj5q9\") pod \"ovn-operator-controller-manager-b6456fdb6-sktq6\" (UID: \"e06aa87c-9f7d-44f7-b96d-27141d2fff93\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sktq6" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.162234 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.165605 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.165798 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-rdq2c" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.176361 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-mnhlt"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.177762 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vhxjt" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.178236 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mnhlt" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.180613 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-mnhlt"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.181708 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-62gb5" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.186917 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.188327 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.189651 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-x5d6x" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.194626 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vtcg5" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.195046 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.202786 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-kvnjv"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.203961 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-kvnjv" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.207874 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-2lk62"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.208745 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2lk62" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.212466 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-8fm8s" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.223494 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-69l55"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.224627 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-xqbrz" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.227577 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-69l55" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.234950 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-vrlzb" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.267691 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn4rx\" (UniqueName: \"kubernetes.io/projected/95ae1fa1-3ac7-4040-86da-150daca97f6b-kube-api-access-gn4rx\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-2b2qk\" (UID: \"95ae1fa1-3ac7-4040-86da-150daca97f6b\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2b2qk" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.267783 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t94cj\" (UniqueName: \"kubernetes.io/projected/310d58eb-6345-4cba-8a01-373c65c7bc40-kube-api-access-t94cj\") pod \"mariadb-operator-controller-manager-56bbcc9d85-45jzr\" (UID: \"310d58eb-6345-4cba-8a01-373c65c7bc40\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-45jzr" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.267819 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whrgb\" (UniqueName: \"kubernetes.io/projected/5f479a93-bb3a-45b8-aa74-60e9f0a7aebd-kube-api-access-whrgb\") pod \"nova-operator-controller-manager-697bc559fc-pk7ws\" (UID: \"5f479a93-bb3a-45b8-aa74-60e9f0a7aebd\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-pk7ws" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.267866 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-778dw\" (UniqueName: \"kubernetes.io/projected/1cfe7f94-e830-4145-a3e7-8a3f694adda1-kube-api-access-778dw\") pod \"octavia-operator-controller-manager-998648c74-kcmvz\" (UID: \"1cfe7f94-e830-4145-a3e7-8a3f694adda1\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-kcmvz" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.267936 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j29nf\" (UniqueName: \"kubernetes.io/projected/2a7907e6-c36f-4534-9460-80c0091af286-kube-api-access-j29nf\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7\" (UID: \"2a7907e6-c36f-4534-9460-80c0091af286\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.267970 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jffl2\" (UniqueName: \"kubernetes.io/projected/66428f64-e34e-4b37-b911-13e7142f1c18-kube-api-access-jffl2\") pod \"manila-operator-controller-manager-7c79b5df47-2spbg\" (UID: \"66428f64-e34e-4b37-b911-13e7142f1c18\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2spbg" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.268025 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7\" (UID: \"2a7907e6-c36f-4534-9460-80c0091af286\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.268046 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj5q9\" (UniqueName: \"kubernetes.io/projected/e06aa87c-9f7d-44f7-b96d-27141d2fff93-kube-api-access-lj5q9\") pod \"ovn-operator-controller-manager-b6456fdb6-sktq6\" (UID: \"e06aa87c-9f7d-44f7-b96d-27141d2fff93\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sktq6" Dec 04 04:10:59 crc kubenswrapper[4806]: E1204 04:10:59.273416 4806 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 04:10:59 crc kubenswrapper[4806]: E1204 04:10:59.273483 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert podName:2a7907e6-c36f-4534-9460-80c0091af286 nodeName:}" failed. No retries permitted until 2025-12-04 04:10:59.773468953 +0000 UTC m=+974.631981881 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" (UID: "2a7907e6-c36f-4534-9460-80c0091af286") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.301437 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-69l55"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.329674 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-2lk62"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.396140 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lx4r\" (UniqueName: \"kubernetes.io/projected/669dd56a-69b0-4315-aacc-c8f085f0b76a-kube-api-access-7lx4r\") pod \"keystone-operator-controller-manager-7765d96ddf-nt5vq\" (UID: \"669dd56a-69b0-4315-aacc-c8f085f0b76a\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nt5vq" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.403051 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t94cj\" (UniqueName: \"kubernetes.io/projected/310d58eb-6345-4cba-8a01-373c65c7bc40-kube-api-access-t94cj\") pod \"mariadb-operator-controller-manager-56bbcc9d85-45jzr\" (UID: \"310d58eb-6345-4cba-8a01-373c65c7bc40\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-45jzr" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.409366 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jffl2\" (UniqueName: \"kubernetes.io/projected/66428f64-e34e-4b37-b911-13e7142f1c18-kube-api-access-jffl2\") pod \"manila-operator-controller-manager-7c79b5df47-2spbg\" (UID: \"66428f64-e34e-4b37-b911-13e7142f1c18\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2spbg" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.417877 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn4rx\" (UniqueName: \"kubernetes.io/projected/95ae1fa1-3ac7-4040-86da-150daca97f6b-kube-api-access-gn4rx\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-2b2qk\" (UID: \"95ae1fa1-3ac7-4040-86da-150daca97f6b\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2b2qk" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.423700 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-kvnjv"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.436853 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whrgb\" (UniqueName: \"kubernetes.io/projected/5f479a93-bb3a-45b8-aa74-60e9f0a7aebd-kube-api-access-whrgb\") pod \"nova-operator-controller-manager-697bc559fc-pk7ws\" (UID: \"5f479a93-bb3a-45b8-aa74-60e9f0a7aebd\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-pk7ws" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.439381 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j29nf\" (UniqueName: \"kubernetes.io/projected/2a7907e6-c36f-4534-9460-80c0091af286-kube-api-access-j29nf\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7\" (UID: \"2a7907e6-c36f-4534-9460-80c0091af286\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.464011 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj5q9\" (UniqueName: \"kubernetes.io/projected/e06aa87c-9f7d-44f7-b96d-27141d2fff93-kube-api-access-lj5q9\") pod \"ovn-operator-controller-manager-b6456fdb6-sktq6\" (UID: \"e06aa87c-9f7d-44f7-b96d-27141d2fff93\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sktq6" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.486775 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p5w4\" (UniqueName: \"kubernetes.io/projected/45192fbb-caf2-4df8-82a6-c222f73b033c-kube-api-access-5p5w4\") pod \"placement-operator-controller-manager-78f8948974-mnhlt\" (UID: \"45192fbb-caf2-4df8-82a6-c222f73b033c\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-mnhlt" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.487215 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-45jzr" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.494990 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf74m\" (UniqueName: \"kubernetes.io/projected/a3827706-d53c-48f3-b4cf-1c240c0f868c-kube-api-access-xf74m\") pod \"watcher-operator-controller-manager-769dc69bc-2lk62\" (UID: \"a3827706-d53c-48f3-b4cf-1c240c0f868c\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2lk62" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.497584 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert\") pod \"infra-operator-controller-manager-57548d458d-g6svq\" (UID: \"04d2c949-c17a-4299-a9be-b56a7c34866c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.497747 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klckh\" (UniqueName: \"kubernetes.io/projected/3580e441-1df2-4fe9-8969-e1621fc554dc-kube-api-access-klckh\") pod \"swift-operator-controller-manager-5f8c65bbfc-fdqbx\" (UID: \"3580e441-1df2-4fe9-8969-e1621fc554dc\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.497873 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnwvp\" (UniqueName: \"kubernetes.io/projected/647621de-6ecf-4ebe-96f2-e85f26e4917e-kube-api-access-xnwvp\") pod \"telemetry-operator-controller-manager-76cc84c6bb-kvnjv\" (UID: \"647621de-6ecf-4ebe-96f2-e85f26e4917e\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-kvnjv" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.497942 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4x79\" (UniqueName: \"kubernetes.io/projected/2473e4f1-90f0-43a0-9e72-0c761d335d76-kube-api-access-l4x79\") pod \"test-operator-controller-manager-5854674fcc-69l55\" (UID: \"2473e4f1-90f0-43a0-9e72-0c761d335d76\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-69l55" Dec 04 04:10:59 crc kubenswrapper[4806]: E1204 04:10:59.499044 4806 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 04:10:59 crc kubenswrapper[4806]: E1204 04:10:59.499127 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert podName:04d2c949-c17a-4299-a9be-b56a7c34866c nodeName:}" failed. No retries permitted until 2025-12-04 04:11:00.499106795 +0000 UTC m=+975.357619723 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert") pod "infra-operator-controller-manager-57548d458d-g6svq" (UID: "04d2c949-c17a-4299-a9be-b56a7c34866c") : secret "infra-operator-webhook-server-cert" not found Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.506236 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-778dw\" (UniqueName: \"kubernetes.io/projected/1cfe7f94-e830-4145-a3e7-8a3f694adda1-kube-api-access-778dw\") pod \"octavia-operator-controller-manager-998648c74-kcmvz\" (UID: \"1cfe7f94-e830-4145-a3e7-8a3f694adda1\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-kcmvz" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.507072 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2b2qk" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.526075 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-pk7ws" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.655381 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kcmvz" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.657137 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klckh\" (UniqueName: \"kubernetes.io/projected/3580e441-1df2-4fe9-8969-e1621fc554dc-kube-api-access-klckh\") pod \"swift-operator-controller-manager-5f8c65bbfc-fdqbx\" (UID: \"3580e441-1df2-4fe9-8969-e1621fc554dc\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.658552 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnwvp\" (UniqueName: \"kubernetes.io/projected/647621de-6ecf-4ebe-96f2-e85f26e4917e-kube-api-access-xnwvp\") pod \"telemetry-operator-controller-manager-76cc84c6bb-kvnjv\" (UID: \"647621de-6ecf-4ebe-96f2-e85f26e4917e\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-kvnjv" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.658605 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4x79\" (UniqueName: \"kubernetes.io/projected/2473e4f1-90f0-43a0-9e72-0c761d335d76-kube-api-access-l4x79\") pod \"test-operator-controller-manager-5854674fcc-69l55\" (UID: \"2473e4f1-90f0-43a0-9e72-0c761d335d76\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-69l55" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.658637 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p5w4\" (UniqueName: \"kubernetes.io/projected/45192fbb-caf2-4df8-82a6-c222f73b033c-kube-api-access-5p5w4\") pod \"placement-operator-controller-manager-78f8948974-mnhlt\" (UID: \"45192fbb-caf2-4df8-82a6-c222f73b033c\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-mnhlt" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.658730 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf74m\" (UniqueName: \"kubernetes.io/projected/a3827706-d53c-48f3-b4cf-1c240c0f868c-kube-api-access-xf74m\") pod \"watcher-operator-controller-manager-769dc69bc-2lk62\" (UID: \"a3827706-d53c-48f3-b4cf-1c240c0f868c\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2lk62" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.660713 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nt5vq" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.667497 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2spbg" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.715865 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4x79\" (UniqueName: \"kubernetes.io/projected/2473e4f1-90f0-43a0-9e72-0c761d335d76-kube-api-access-l4x79\") pod \"test-operator-controller-manager-5854674fcc-69l55\" (UID: \"2473e4f1-90f0-43a0-9e72-0c761d335d76\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-69l55" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.716218 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sktq6" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.749225 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.752751 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klckh\" (UniqueName: \"kubernetes.io/projected/3580e441-1df2-4fe9-8969-e1621fc554dc-kube-api-access-klckh\") pod \"swift-operator-controller-manager-5f8c65bbfc-fdqbx\" (UID: \"3580e441-1df2-4fe9-8969-e1621fc554dc\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.753966 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnwvp\" (UniqueName: \"kubernetes.io/projected/647621de-6ecf-4ebe-96f2-e85f26e4917e-kube-api-access-xnwvp\") pod \"telemetry-operator-controller-manager-76cc84c6bb-kvnjv\" (UID: \"647621de-6ecf-4ebe-96f2-e85f26e4917e\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-kvnjv" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.756849 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2"] Dec 04 04:10:59 crc kubenswrapper[4806]: E1204 04:10:59.757196 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cb9a7f5-2b1d-436d-9c56-b0898279ba66" containerName="extract-utilities" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.757210 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cb9a7f5-2b1d-436d-9c56-b0898279ba66" containerName="extract-utilities" Dec 04 04:10:59 crc kubenswrapper[4806]: E1204 04:10:59.757225 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cb9a7f5-2b1d-436d-9c56-b0898279ba66" containerName="registry-server" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.757239 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cb9a7f5-2b1d-436d-9c56-b0898279ba66" containerName="registry-server" Dec 04 04:10:59 crc kubenswrapper[4806]: E1204 04:10:59.757251 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cb9a7f5-2b1d-436d-9c56-b0898279ba66" containerName="extract-content" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.757258 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cb9a7f5-2b1d-436d-9c56-b0898279ba66" containerName="extract-content" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.757395 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cb9a7f5-2b1d-436d-9c56-b0898279ba66" containerName="registry-server" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.757990 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.770232 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-utilities\") pod \"8cb9a7f5-2b1d-436d-9c56-b0898279ba66\" (UID: \"8cb9a7f5-2b1d-436d-9c56-b0898279ba66\") " Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.772075 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lf6r\" (UniqueName: \"kubernetes.io/projected/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-kube-api-access-4lf6r\") pod \"8cb9a7f5-2b1d-436d-9c56-b0898279ba66\" (UID: \"8cb9a7f5-2b1d-436d-9c56-b0898279ba66\") " Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.772127 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-catalog-content\") pod \"8cb9a7f5-2b1d-436d-9c56-b0898279ba66\" (UID: \"8cb9a7f5-2b1d-436d-9c56-b0898279ba66\") " Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.772378 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp5f6\" (UniqueName: \"kubernetes.io/projected/ff1a6781-908b-419f-8a02-10abbbfa0816-kube-api-access-xp5f6\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.772520 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.772565 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.784061 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-utilities" (OuterVolumeSpecName: "utilities") pod "8cb9a7f5-2b1d-436d-9c56-b0898279ba66" (UID: "8cb9a7f5-2b1d-436d-9c56-b0898279ba66"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.784106 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.796550 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf74m\" (UniqueName: \"kubernetes.io/projected/a3827706-d53c-48f3-b4cf-1c240c0f868c-kube-api-access-xf74m\") pod \"watcher-operator-controller-manager-769dc69bc-2lk62\" (UID: \"a3827706-d53c-48f3-b4cf-1c240c0f868c\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2lk62" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.796901 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.797026 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.797106 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-9srkm" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.797166 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p5w4\" (UniqueName: \"kubernetes.io/projected/45192fbb-caf2-4df8-82a6-c222f73b033c-kube-api-access-5p5w4\") pod \"placement-operator-controller-manager-78f8948974-mnhlt\" (UID: \"45192fbb-caf2-4df8-82a6-c222f73b033c\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-mnhlt" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.803143 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-kube-api-access-4lf6r" (OuterVolumeSpecName: "kube-api-access-4lf6r") pod "8cb9a7f5-2b1d-436d-9c56-b0898279ba66" (UID: "8cb9a7f5-2b1d-436d-9c56-b0898279ba66"). InnerVolumeSpecName "kube-api-access-4lf6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.811305 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.876037 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp5f6\" (UniqueName: \"kubernetes.io/projected/ff1a6781-908b-419f-8a02-10abbbfa0816-kube-api-access-xp5f6\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.876097 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7\" (UID: \"2a7907e6-c36f-4534-9460-80c0091af286\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.876133 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.876162 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.876227 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lf6r\" (UniqueName: \"kubernetes.io/projected/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-kube-api-access-4lf6r\") on node \"crc\" DevicePath \"\"" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.876238 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:10:59 crc kubenswrapper[4806]: E1204 04:10:59.876321 4806 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 04:10:59 crc kubenswrapper[4806]: E1204 04:10:59.876366 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs podName:ff1a6781-908b-419f-8a02-10abbbfa0816 nodeName:}" failed. No retries permitted until 2025-12-04 04:11:00.376351721 +0000 UTC m=+975.234864649 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs") pod "openstack-operator-controller-manager-8d698776b-npcz2" (UID: "ff1a6781-908b-419f-8a02-10abbbfa0816") : secret "webhook-server-cert" not found Dec 04 04:10:59 crc kubenswrapper[4806]: E1204 04:10:59.876792 4806 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 04:10:59 crc kubenswrapper[4806]: E1204 04:10:59.876818 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert podName:2a7907e6-c36f-4534-9460-80c0091af286 nodeName:}" failed. No retries permitted until 2025-12-04 04:11:00.876810525 +0000 UTC m=+975.735323453 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" (UID: "2a7907e6-c36f-4534-9460-80c0091af286") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 04:10:59 crc kubenswrapper[4806]: E1204 04:10:59.876851 4806 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 04:10:59 crc kubenswrapper[4806]: E1204 04:10:59.876873 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs podName:ff1a6781-908b-419f-8a02-10abbbfa0816 nodeName:}" failed. No retries permitted until 2025-12-04 04:11:00.376865976 +0000 UTC m=+975.235378904 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs") pod "openstack-operator-controller-manager-8d698776b-npcz2" (UID: "ff1a6781-908b-419f-8a02-10abbbfa0816") : secret "metrics-server-cert" not found Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.930469 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-g2dfw"] Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.934407 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp5f6\" (UniqueName: \"kubernetes.io/projected/ff1a6781-908b-419f-8a02-10abbbfa0816-kube-api-access-xp5f6\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.938371 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8cb9a7f5-2b1d-436d-9c56-b0898279ba66" (UID: "8cb9a7f5-2b1d-436d-9c56-b0898279ba66"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.942149 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tpq2v" event={"ID":"8cb9a7f5-2b1d-436d-9c56-b0898279ba66","Type":"ContainerDied","Data":"82cef718e3ba375835bc15f8250fc69c841e8b299151a39e391dd1b72f925d4e"} Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.942197 4806 scope.go:117] "RemoveContainer" containerID="f925a91b8b14fa816810bc818f02263714cea10c2802c9de63748c696a514385" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.942313 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tpq2v" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.970467 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2lk62" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.979519 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cb9a7f5-2b1d-436d-9c56-b0898279ba66-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:10:59 crc kubenswrapper[4806]: I1204 04:10:59.979740 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-kvnjv" Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.033099 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-69l55" Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.033667 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cwnth"] Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.034694 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cwnth" Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.043498 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-shl7h" Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.056680 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cwnth"] Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.086774 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52snr\" (UniqueName: \"kubernetes.io/projected/be179b75-6e57-4537-b3ee-a57dc049948e-kube-api-access-52snr\") pod \"rabbitmq-cluster-operator-manager-668c99d594-cwnth\" (UID: \"be179b75-6e57-4537-b3ee-a57dc049948e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cwnth" Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.087081 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mnhlt" Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.172006 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-xd6pj"] Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.190758 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52snr\" (UniqueName: \"kubernetes.io/projected/be179b75-6e57-4537-b3ee-a57dc049948e-kube-api-access-52snr\") pod \"rabbitmq-cluster-operator-manager-668c99d594-cwnth\" (UID: \"be179b75-6e57-4537-b3ee-a57dc049948e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cwnth" Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.233687 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52snr\" (UniqueName: \"kubernetes.io/projected/be179b75-6e57-4537-b3ee-a57dc049948e-kube-api-access-52snr\") pod \"rabbitmq-cluster-operator-manager-668c99d594-cwnth\" (UID: \"be179b75-6e57-4537-b3ee-a57dc049948e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cwnth" Dec 04 04:11:00 crc kubenswrapper[4806]: W1204 04:11:00.288980 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda022d699_4a2b_49e2_8d8f_787a3d888df8.slice/crio-93de371f8d5f5662659d11de62c3e869c79e6656f88512f10e9598c3f6f74a3a WatchSource:0}: Error finding container 93de371f8d5f5662659d11de62c3e869c79e6656f88512f10e9598c3f6f74a3a: Status 404 returned error can't find the container with id 93de371f8d5f5662659d11de62c3e869c79e6656f88512f10e9598c3f6f74a3a Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.341814 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-25nxx"] Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.356786 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tpq2v"] Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.358310 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cwnth" Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.360579 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tpq2v"] Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.380278 4806 scope.go:117] "RemoveContainer" containerID="b1cbad48e2334478d378dc4810cfd11f38ed6364195e9f9c5168e23ecf12708e" Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.395898 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:11:00 crc kubenswrapper[4806]: E1204 04:11:00.396415 4806 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 04:11:00 crc kubenswrapper[4806]: E1204 04:11:00.396474 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs podName:ff1a6781-908b-419f-8a02-10abbbfa0816 nodeName:}" failed. No retries permitted until 2025-12-04 04:11:01.396457924 +0000 UTC m=+976.254970842 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs") pod "openstack-operator-controller-manager-8d698776b-npcz2" (UID: "ff1a6781-908b-419f-8a02-10abbbfa0816") : secret "metrics-server-cert" not found Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.397976 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:11:00 crc kubenswrapper[4806]: E1204 04:11:00.398150 4806 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 04:11:00 crc kubenswrapper[4806]: E1204 04:11:00.398191 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs podName:ff1a6781-908b-419f-8a02-10abbbfa0816 nodeName:}" failed. No retries permitted until 2025-12-04 04:11:01.398176858 +0000 UTC m=+976.256689786 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs") pod "openstack-operator-controller-manager-8d698776b-npcz2" (UID: "ff1a6781-908b-419f-8a02-10abbbfa0816") : secret "webhook-server-cert" not found Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.604847 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert\") pod \"infra-operator-controller-manager-57548d458d-g6svq\" (UID: \"04d2c949-c17a-4299-a9be-b56a7c34866c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" Dec 04 04:11:00 crc kubenswrapper[4806]: E1204 04:11:00.605308 4806 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 04:11:00 crc kubenswrapper[4806]: E1204 04:11:00.605356 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert podName:04d2c949-c17a-4299-a9be-b56a7c34866c nodeName:}" failed. No retries permitted until 2025-12-04 04:11:02.605341791 +0000 UTC m=+977.463854719 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert") pod "infra-operator-controller-manager-57548d458d-g6svq" (UID: "04d2c949-c17a-4299-a9be-b56a7c34866c") : secret "infra-operator-webhook-server-cert" not found Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.620403 4806 scope.go:117] "RemoveContainer" containerID="79dbef329e4746d2879807f0eb67ae2fc37b6b7dc74f4b102ea6c7e58c6507e9" Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.778319 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-mxmf2"] Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.790422 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-f6k5f"] Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.854555 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vtcg5"] Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.873947 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vhxjt"] Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.910633 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7\" (UID: \"2a7907e6-c36f-4534-9460-80c0091af286\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" Dec 04 04:11:00 crc kubenswrapper[4806]: E1204 04:11:00.911048 4806 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 04:11:00 crc kubenswrapper[4806]: E1204 04:11:00.911095 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert podName:2a7907e6-c36f-4534-9460-80c0091af286 nodeName:}" failed. No retries permitted until 2025-12-04 04:11:02.911082275 +0000 UTC m=+977.769595203 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" (UID: "2a7907e6-c36f-4534-9460-80c0091af286") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 04:11:00 crc kubenswrapper[4806]: W1204 04:11:00.916396 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod084c526b_94d8_43ae_9ffb_33aa215de8c2.slice/crio-18b60f637e4805a7cc9f0960821f2b0a5d80311e81903898e62dde2a9c81fc80 WatchSource:0}: Error finding container 18b60f637e4805a7cc9f0960821f2b0a5d80311e81903898e62dde2a9c81fc80: Status 404 returned error can't find the container with id 18b60f637e4805a7cc9f0960821f2b0a5d80311e81903898e62dde2a9c81fc80 Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.957579 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vtcg5" event={"ID":"084c526b-94d8-43ae-9ffb-33aa215de8c2","Type":"ContainerStarted","Data":"18b60f637e4805a7cc9f0960821f2b0a5d80311e81903898e62dde2a9c81fc80"} Dec 04 04:11:00 crc kubenswrapper[4806]: I1204 04:11:00.968392 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mxmf2" event={"ID":"91ab6e01-027e-4e7a-a250-374b4e8873fc","Type":"ContainerStarted","Data":"e63963f328271bf7932ee17cabd08deab2eec2f0d3c97f867b7f7e40f95820ec"} Dec 04 04:11:01 crc kubenswrapper[4806]: I1204 04:11:01.032523 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-xd6pj" event={"ID":"7213e0bc-21af-4df7-929d-f915e343f184","Type":"ContainerStarted","Data":"75e366ff590d8f91107b60c249f276b2c6686ebcb53b6df60986b19cc29619d4"} Dec 04 04:11:01 crc kubenswrapper[4806]: I1204 04:11:01.051780 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g2dfw" event={"ID":"b2bba37e-d570-432f-9520-9b5de2005e5c","Type":"ContainerStarted","Data":"d28a9d96906eaf3ff8b246cb57cf78337d36616f523257ab4fbf59e5808d3b7e"} Dec 04 04:11:01 crc kubenswrapper[4806]: I1204 04:11:01.073480 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-25nxx" event={"ID":"a022d699-4a2b-49e2-8d8f-787a3d888df8","Type":"ContainerStarted","Data":"93de371f8d5f5662659d11de62c3e869c79e6656f88512f10e9598c3f6f74a3a"} Dec 04 04:11:01 crc kubenswrapper[4806]: I1204 04:11:01.434540 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:11:01 crc kubenswrapper[4806]: I1204 04:11:01.435096 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:11:01 crc kubenswrapper[4806]: E1204 04:11:01.437213 4806 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 04:11:01 crc kubenswrapper[4806]: E1204 04:11:01.437320 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs podName:ff1a6781-908b-419f-8a02-10abbbfa0816 nodeName:}" failed. No retries permitted until 2025-12-04 04:11:03.437286681 +0000 UTC m=+978.295799609 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs") pod "openstack-operator-controller-manager-8d698776b-npcz2" (UID: "ff1a6781-908b-419f-8a02-10abbbfa0816") : secret "metrics-server-cert" not found Dec 04 04:11:01 crc kubenswrapper[4806]: E1204 04:11:01.463222 4806 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 04:11:01 crc kubenswrapper[4806]: E1204 04:11:01.463319 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs podName:ff1a6781-908b-419f-8a02-10abbbfa0816 nodeName:}" failed. No retries permitted until 2025-12-04 04:11:03.463296113 +0000 UTC m=+978.321809041 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs") pod "openstack-operator-controller-manager-8d698776b-npcz2" (UID: "ff1a6781-908b-419f-8a02-10abbbfa0816") : secret "webhook-server-cert" not found Dec 04 04:11:01 crc kubenswrapper[4806]: I1204 04:11:01.485011 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cb9a7f5-2b1d-436d-9c56-b0898279ba66" path="/var/lib/kubelet/pods/8cb9a7f5-2b1d-436d-9c56-b0898279ba66/volumes" Dec 04 04:11:01 crc kubenswrapper[4806]: I1204 04:11:01.542974 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-45jzr"] Dec 04 04:11:01 crc kubenswrapper[4806]: I1204 04:11:01.770344 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-mnhlt"] Dec 04 04:11:01 crc kubenswrapper[4806]: W1204 04:11:01.776412 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45192fbb_caf2_4df8_82a6_c222f73b033c.slice/crio-948966515d345676deb2319a3fb155cff2b3622b19f08d6f2c0bd8107b54c913 WatchSource:0}: Error finding container 948966515d345676deb2319a3fb155cff2b3622b19f08d6f2c0bd8107b54c913: Status 404 returned error can't find the container with id 948966515d345676deb2319a3fb155cff2b3622b19f08d6f2c0bd8107b54c913 Dec 04 04:11:01 crc kubenswrapper[4806]: W1204 04:11:01.784608 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cfe7f94_e830_4145_a3e7_8a3f694adda1.slice/crio-8f211666ea68909a29e4309867595c113a543f84f52d34e3193352b6f33de346 WatchSource:0}: Error finding container 8f211666ea68909a29e4309867595c113a543f84f52d34e3193352b6f33de346: Status 404 returned error can't find the container with id 8f211666ea68909a29e4309867595c113a543f84f52d34e3193352b6f33de346 Dec 04 04:11:01 crc kubenswrapper[4806]: I1204 04:11:01.790968 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-kcmvz"] Dec 04 04:11:01 crc kubenswrapper[4806]: I1204 04:11:01.865278 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-2spbg"] Dec 04 04:11:01 crc kubenswrapper[4806]: I1204 04:11:01.957100 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-kvnjv"] Dec 04 04:11:01 crc kubenswrapper[4806]: I1204 04:11:01.964172 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cwnth"] Dec 04 04:11:01 crc kubenswrapper[4806]: I1204 04:11:01.977697 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-nt5vq"] Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.016859 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-pk7ws"] Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.026150 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2b2qk"] Dec 04 04:11:02 crc kubenswrapper[4806]: W1204 04:11:02.085002 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f479a93_bb3a_45b8_aa74_60e9f0a7aebd.slice/crio-6416cf09126d99ca59cf0e7fa62f82c78504420206ab9d67d8c78337844bd965 WatchSource:0}: Error finding container 6416cf09126d99ca59cf0e7fa62f82c78504420206ab9d67d8c78337844bd965: Status 404 returned error can't find the container with id 6416cf09126d99ca59cf0e7fa62f82c78504420206ab9d67d8c78337844bd965 Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.094249 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx"] Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.107051 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-sktq6"] Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.113940 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-69l55"] Dec 04 04:11:02 crc kubenswrapper[4806]: W1204 04:11:02.126088 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95ae1fa1_3ac7_4040_86da_150daca97f6b.slice/crio-87d160c6d9b8cdef03fa2bbfaba971c1fa1b0db77d26360a9062d874d86d89c5 WatchSource:0}: Error finding container 87d160c6d9b8cdef03fa2bbfaba971c1fa1b0db77d26360a9062d874d86d89c5: Status 404 returned error can't find the container with id 87d160c6d9b8cdef03fa2bbfaba971c1fa1b0db77d26360a9062d874d86d89c5 Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.131002 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-2lk62"] Dec 04 04:11:02 crc kubenswrapper[4806]: W1204 04:11:02.147742 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode06aa87c_9f7d_44f7_b96d_27141d2fff93.slice/crio-ce6b298bcb3ce1b0f516b1d541ea415884da77007165116881c57d6da458ac40 WatchSource:0}: Error finding container ce6b298bcb3ce1b0f516b1d541ea415884da77007165116881c57d6da458ac40: Status 404 returned error can't find the container with id ce6b298bcb3ce1b0f516b1d541ea415884da77007165116881c57d6da458ac40 Dec 04 04:11:02 crc kubenswrapper[4806]: W1204 04:11:02.152534 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3580e441_1df2_4fe9_8969_e1621fc554dc.slice/crio-7b198b781489addae1db6bae7ab32bf89b0d3cf91e768ddbcc6eab537c89b3c5 WatchSource:0}: Error finding container 7b198b781489addae1db6bae7ab32bf89b0d3cf91e768ddbcc6eab537c89b3c5: Status 404 returned error can't find the container with id 7b198b781489addae1db6bae7ab32bf89b0d3cf91e768ddbcc6eab537c89b3c5 Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.160516 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2spbg" event={"ID":"66428f64-e34e-4b37-b911-13e7142f1c18","Type":"ContainerStarted","Data":"f4d61af53b678ac740c11f06adeeab78f1d19d9129c55ba8c4ba71e612befff7"} Dec 04 04:11:02 crc kubenswrapper[4806]: E1204 04:11:02.160956 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-klckh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-fdqbx_openstack-operators(3580e441-1df2-4fe9-8969-e1621fc554dc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 04:11:02 crc kubenswrapper[4806]: E1204 04:11:02.163437 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-klckh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-fdqbx_openstack-operators(3580e441-1df2-4fe9-8969-e1621fc554dc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 04:11:02 crc kubenswrapper[4806]: E1204 04:11:02.164776 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx" podUID="3580e441-1df2-4fe9-8969-e1621fc554dc" Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.166946 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cwnth" event={"ID":"be179b75-6e57-4537-b3ee-a57dc049948e","Type":"ContainerStarted","Data":"ecd760849d9beb138f26c51a582ab95572fd569d8533a5213351bcd4247a3ee5"} Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.170199 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mnhlt" event={"ID":"45192fbb-caf2-4df8-82a6-c222f73b033c","Type":"ContainerStarted","Data":"948966515d345676deb2319a3fb155cff2b3622b19f08d6f2c0bd8107b54c913"} Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.173982 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-f6k5f" event={"ID":"5f55ddcf-ca5d-4249-9384-e541ae7d4536","Type":"ContainerStarted","Data":"92293bc86737166733781d7c75f08f4e61395c33465b3af4c09db8feada5cc18"} Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.175998 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nt5vq" event={"ID":"669dd56a-69b0-4315-aacc-c8f085f0b76a","Type":"ContainerStarted","Data":"32c109b85d83f09422296b63f24e14366b2c79acddc92ec4e7dc9bba1105c847"} Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.184680 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vhxjt" event={"ID":"b496f26d-2493-416f-a0f8-5b33e41dbda0","Type":"ContainerStarted","Data":"800658a4426a6ba4075e10323dd0a9db6a701abf1ef82d4b88b73ebf89bd7366"} Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.189132 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-45jzr" event={"ID":"310d58eb-6345-4cba-8a01-373c65c7bc40","Type":"ContainerStarted","Data":"a6e2f60c37e85164aeb0d2e20c644cd3dcfc993992346a4c0ae2677dc50b93bd"} Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.201577 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-kvnjv" event={"ID":"647621de-6ecf-4ebe-96f2-e85f26e4917e","Type":"ContainerStarted","Data":"e2d93a0b2c89c559efdba78f8ff4555d08cc80dd015ce9dc6bff9cd585587c7d"} Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.203325 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kcmvz" event={"ID":"1cfe7f94-e830-4145-a3e7-8a3f694adda1","Type":"ContainerStarted","Data":"8f211666ea68909a29e4309867595c113a543f84f52d34e3193352b6f33de346"} Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.666853 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert\") pod \"infra-operator-controller-manager-57548d458d-g6svq\" (UID: \"04d2c949-c17a-4299-a9be-b56a7c34866c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" Dec 04 04:11:02 crc kubenswrapper[4806]: E1204 04:11:02.667039 4806 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 04:11:02 crc kubenswrapper[4806]: E1204 04:11:02.667109 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert podName:04d2c949-c17a-4299-a9be-b56a7c34866c nodeName:}" failed. No retries permitted until 2025-12-04 04:11:06.66709038 +0000 UTC m=+981.525603308 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert") pod "infra-operator-controller-manager-57548d458d-g6svq" (UID: "04d2c949-c17a-4299-a9be-b56a7c34866c") : secret "infra-operator-webhook-server-cert" not found Dec 04 04:11:02 crc kubenswrapper[4806]: I1204 04:11:02.983679 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7\" (UID: \"2a7907e6-c36f-4534-9460-80c0091af286\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" Dec 04 04:11:02 crc kubenswrapper[4806]: E1204 04:11:02.984069 4806 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 04:11:02 crc kubenswrapper[4806]: E1204 04:11:02.984119 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert podName:2a7907e6-c36f-4534-9460-80c0091af286 nodeName:}" failed. No retries permitted until 2025-12-04 04:11:06.984105203 +0000 UTC m=+981.842618131 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" (UID: "2a7907e6-c36f-4534-9460-80c0091af286") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 04:11:03 crc kubenswrapper[4806]: I1204 04:11:03.234073 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2b2qk" event={"ID":"95ae1fa1-3ac7-4040-86da-150daca97f6b","Type":"ContainerStarted","Data":"87d160c6d9b8cdef03fa2bbfaba971c1fa1b0db77d26360a9062d874d86d89c5"} Dec 04 04:11:03 crc kubenswrapper[4806]: I1204 04:11:03.239329 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx" event={"ID":"3580e441-1df2-4fe9-8969-e1621fc554dc","Type":"ContainerStarted","Data":"7b198b781489addae1db6bae7ab32bf89b0d3cf91e768ddbcc6eab537c89b3c5"} Dec 04 04:11:03 crc kubenswrapper[4806]: I1204 04:11:03.264486 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2lk62" event={"ID":"a3827706-d53c-48f3-b4cf-1c240c0f868c","Type":"ContainerStarted","Data":"329acd5cbe03c74a9398852903a75ad4f8cf2229f1d6ae2af999ce4771207820"} Dec 04 04:11:03 crc kubenswrapper[4806]: E1204 04:11:03.264647 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx" podUID="3580e441-1df2-4fe9-8969-e1621fc554dc" Dec 04 04:11:03 crc kubenswrapper[4806]: I1204 04:11:03.270414 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sktq6" event={"ID":"e06aa87c-9f7d-44f7-b96d-27141d2fff93","Type":"ContainerStarted","Data":"ce6b298bcb3ce1b0f516b1d541ea415884da77007165116881c57d6da458ac40"} Dec 04 04:11:03 crc kubenswrapper[4806]: I1204 04:11:03.287957 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-pk7ws" event={"ID":"5f479a93-bb3a-45b8-aa74-60e9f0a7aebd","Type":"ContainerStarted","Data":"6416cf09126d99ca59cf0e7fa62f82c78504420206ab9d67d8c78337844bd965"} Dec 04 04:11:03 crc kubenswrapper[4806]: I1204 04:11:03.293477 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-69l55" event={"ID":"2473e4f1-90f0-43a0-9e72-0c761d335d76","Type":"ContainerStarted","Data":"484f84522ac2de54d532f3622095f2d6a7ab958a9b7f4f887703b7b61bb908e0"} Dec 04 04:11:03 crc kubenswrapper[4806]: I1204 04:11:03.443679 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:11:03 crc kubenswrapper[4806]: E1204 04:11:03.443863 4806 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 04:11:03 crc kubenswrapper[4806]: E1204 04:11:03.443912 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs podName:ff1a6781-908b-419f-8a02-10abbbfa0816 nodeName:}" failed. No retries permitted until 2025-12-04 04:11:07.443897521 +0000 UTC m=+982.302410439 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs") pod "openstack-operator-controller-manager-8d698776b-npcz2" (UID: "ff1a6781-908b-419f-8a02-10abbbfa0816") : secret "metrics-server-cert" not found Dec 04 04:11:03 crc kubenswrapper[4806]: I1204 04:11:03.552229 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:11:03 crc kubenswrapper[4806]: E1204 04:11:03.552400 4806 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 04:11:03 crc kubenswrapper[4806]: E1204 04:11:03.552467 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs podName:ff1a6781-908b-419f-8a02-10abbbfa0816 nodeName:}" failed. No retries permitted until 2025-12-04 04:11:07.552447687 +0000 UTC m=+982.410960615 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs") pod "openstack-operator-controller-manager-8d698776b-npcz2" (UID: "ff1a6781-908b-419f-8a02-10abbbfa0816") : secret "webhook-server-cert" not found Dec 04 04:11:04 crc kubenswrapper[4806]: E1204 04:11:04.307851 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx" podUID="3580e441-1df2-4fe9-8969-e1621fc554dc" Dec 04 04:11:06 crc kubenswrapper[4806]: I1204 04:11:06.721717 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert\") pod \"infra-operator-controller-manager-57548d458d-g6svq\" (UID: \"04d2c949-c17a-4299-a9be-b56a7c34866c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" Dec 04 04:11:06 crc kubenswrapper[4806]: E1204 04:11:06.721953 4806 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 04:11:06 crc kubenswrapper[4806]: E1204 04:11:06.722152 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert podName:04d2c949-c17a-4299-a9be-b56a7c34866c nodeName:}" failed. No retries permitted until 2025-12-04 04:11:14.722131869 +0000 UTC m=+989.580644797 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert") pod "infra-operator-controller-manager-57548d458d-g6svq" (UID: "04d2c949-c17a-4299-a9be-b56a7c34866c") : secret "infra-operator-webhook-server-cert" not found Dec 04 04:11:07 crc kubenswrapper[4806]: I1204 04:11:07.025470 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7\" (UID: \"2a7907e6-c36f-4534-9460-80c0091af286\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" Dec 04 04:11:07 crc kubenswrapper[4806]: E1204 04:11:07.025594 4806 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 04:11:07 crc kubenswrapper[4806]: E1204 04:11:07.025663 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert podName:2a7907e6-c36f-4534-9460-80c0091af286 nodeName:}" failed. No retries permitted until 2025-12-04 04:11:15.02564819 +0000 UTC m=+989.884161118 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" (UID: "2a7907e6-c36f-4534-9460-80c0091af286") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 04:11:07 crc kubenswrapper[4806]: I1204 04:11:07.530867 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:11:07 crc kubenswrapper[4806]: E1204 04:11:07.531075 4806 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 04:11:07 crc kubenswrapper[4806]: E1204 04:11:07.531167 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs podName:ff1a6781-908b-419f-8a02-10abbbfa0816 nodeName:}" failed. No retries permitted until 2025-12-04 04:11:15.531146946 +0000 UTC m=+990.389659884 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs") pod "openstack-operator-controller-manager-8d698776b-npcz2" (UID: "ff1a6781-908b-419f-8a02-10abbbfa0816") : secret "metrics-server-cert" not found Dec 04 04:11:07 crc kubenswrapper[4806]: I1204 04:11:07.632324 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:11:07 crc kubenswrapper[4806]: E1204 04:11:07.632532 4806 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 04:11:07 crc kubenswrapper[4806]: E1204 04:11:07.632645 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs podName:ff1a6781-908b-419f-8a02-10abbbfa0816 nodeName:}" failed. No retries permitted until 2025-12-04 04:11:15.632604199 +0000 UTC m=+990.491117127 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs") pod "openstack-operator-controller-manager-8d698776b-npcz2" (UID: "ff1a6781-908b-419f-8a02-10abbbfa0816") : secret "webhook-server-cert" not found Dec 04 04:11:14 crc kubenswrapper[4806]: I1204 04:11:14.770619 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert\") pod \"infra-operator-controller-manager-57548d458d-g6svq\" (UID: \"04d2c949-c17a-4299-a9be-b56a7c34866c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" Dec 04 04:11:14 crc kubenswrapper[4806]: I1204 04:11:14.790229 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/04d2c949-c17a-4299-a9be-b56a7c34866c-cert\") pod \"infra-operator-controller-manager-57548d458d-g6svq\" (UID: \"04d2c949-c17a-4299-a9be-b56a7c34866c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" Dec 04 04:11:14 crc kubenswrapper[4806]: I1204 04:11:14.932538 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" Dec 04 04:11:15 crc kubenswrapper[4806]: I1204 04:11:15.075421 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7\" (UID: \"2a7907e6-c36f-4534-9460-80c0091af286\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" Dec 04 04:11:15 crc kubenswrapper[4806]: I1204 04:11:15.083818 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2a7907e6-c36f-4534-9460-80c0091af286-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7\" (UID: \"2a7907e6-c36f-4534-9460-80c0091af286\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" Dec 04 04:11:15 crc kubenswrapper[4806]: I1204 04:11:15.336976 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" Dec 04 04:11:15 crc kubenswrapper[4806]: I1204 04:11:15.583969 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:11:15 crc kubenswrapper[4806]: I1204 04:11:15.593843 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-metrics-certs\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:11:15 crc kubenswrapper[4806]: I1204 04:11:15.685807 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:11:15 crc kubenswrapper[4806]: I1204 04:11:15.693637 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ff1a6781-908b-419f-8a02-10abbbfa0816-webhook-certs\") pod \"openstack-operator-controller-manager-8d698776b-npcz2\" (UID: \"ff1a6781-908b-419f-8a02-10abbbfa0816\") " pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:11:15 crc kubenswrapper[4806]: I1204 04:11:15.737405 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:11:16 crc kubenswrapper[4806]: E1204 04:11:16.414216 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85" Dec 04 04:11:16 crc kubenswrapper[4806]: E1204 04:11:16.415613 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wzlmc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-mxmf2_openstack-operators(91ab6e01-027e-4e7a-a250-374b4e8873fc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:11:17 crc kubenswrapper[4806]: E1204 04:11:17.040574 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9" Dec 04 04:11:17 crc kubenswrapper[4806]: E1204 04:11:17.042760 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jffl2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-2spbg_openstack-operators(66428f64-e34e-4b37-b911-13e7142f1c18): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:11:18 crc kubenswrapper[4806]: E1204 04:11:18.506982 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385" Dec 04 04:11:18 crc kubenswrapper[4806]: E1204 04:11:18.507407 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xnwvp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-kvnjv_openstack-operators(647621de-6ecf-4ebe-96f2-e85f26e4917e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:11:19 crc kubenswrapper[4806]: E1204 04:11:19.116349 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 04 04:11:19 crc kubenswrapper[4806]: E1204 04:11:19.116541 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5p5w4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-mnhlt_openstack-operators(45192fbb-caf2-4df8-82a6-c222f73b033c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:11:20 crc kubenswrapper[4806]: E1204 04:11:20.852971 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 04 04:11:20 crc kubenswrapper[4806]: E1204 04:11:20.854318 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-778dw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-kcmvz_openstack-operators(1cfe7f94-e830-4145-a3e7-8a3f694adda1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:11:20 crc kubenswrapper[4806]: I1204 04:11:20.862315 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 04:11:23 crc kubenswrapper[4806]: E1204 04:11:23.968095 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 04 04:11:23 crc kubenswrapper[4806]: E1204 04:11:23.968645 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gn4rx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-2b2qk_openstack-operators(95ae1fa1-3ac7-4040-86da-150daca97f6b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:11:26 crc kubenswrapper[4806]: E1204 04:11:26.272315 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621" Dec 04 04:11:26 crc kubenswrapper[4806]: E1204 04:11:26.272704 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xf74m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-2lk62_openstack-operators(a3827706-d53c-48f3-b4cf-1c240c0f868c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:11:28 crc kubenswrapper[4806]: E1204 04:11:28.241325 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801" Dec 04 04:11:28 crc kubenswrapper[4806]: E1204 04:11:28.241881 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gh48x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-25nxx_openstack-operators(a022d699-4a2b-49e2-8d8f-787a3d888df8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:11:28 crc kubenswrapper[4806]: E1204 04:11:28.922690 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809" Dec 04 04:11:28 crc kubenswrapper[4806]: E1204 04:11:28.922909 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cddnt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-xd6pj_openstack-operators(7213e0bc-21af-4df7-929d-f915e343f184): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:11:29 crc kubenswrapper[4806]: E1204 04:11:29.487770 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7" Dec 04 04:11:29 crc kubenswrapper[4806]: E1204 04:11:29.488105 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t94cj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-45jzr_openstack-operators(310d58eb-6345-4cba-8a01-373c65c7bc40): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:11:30 crc kubenswrapper[4806]: E1204 04:11:30.046822 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530" Dec 04 04:11:30 crc kubenswrapper[4806]: E1204 04:11:30.047460 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ttnjh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-f6k5f_openstack-operators(5f55ddcf-ca5d-4249-9384-e541ae7d4536): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:11:30 crc kubenswrapper[4806]: E1204 04:11:30.703751 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 04 04:11:30 crc kubenswrapper[4806]: E1204 04:11:30.703992 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-52snr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-cwnth_openstack-operators(be179b75-6e57-4537-b3ee-a57dc049948e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:11:30 crc kubenswrapper[4806]: E1204 04:11:30.705184 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cwnth" podUID="be179b75-6e57-4537-b3ee-a57dc049948e" Dec 04 04:11:31 crc kubenswrapper[4806]: E1204 04:11:31.322700 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 04 04:11:31 crc kubenswrapper[4806]: E1204 04:11:31.322899 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7lx4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-nt5vq_openstack-operators(669dd56a-69b0-4315-aacc-c8f085f0b76a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:11:31 crc kubenswrapper[4806]: E1204 04:11:31.577044 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cwnth" podUID="be179b75-6e57-4537-b3ee-a57dc049948e" Dec 04 04:11:32 crc kubenswrapper[4806]: E1204 04:11:32.250872 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 04 04:11:32 crc kubenswrapper[4806]: E1204 04:11:32.251108 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-whrgb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-pk7ws_openstack-operators(5f479a93-bb3a-45b8-aa74-60e9f0a7aebd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:11:32 crc kubenswrapper[4806]: I1204 04:11:32.868315 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-g6svq"] Dec 04 04:11:32 crc kubenswrapper[4806]: I1204 04:11:32.891963 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2"] Dec 04 04:11:32 crc kubenswrapper[4806]: I1204 04:11:32.976264 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7"] Dec 04 04:11:33 crc kubenswrapper[4806]: W1204 04:11:33.223536 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04d2c949_c17a_4299_a9be_b56a7c34866c.slice/crio-17c01f21b0522d6e3f2c9195bdb943e806442f051d7550aa2578f5c41aa803af WatchSource:0}: Error finding container 17c01f21b0522d6e3f2c9195bdb943e806442f051d7550aa2578f5c41aa803af: Status 404 returned error can't find the container with id 17c01f21b0522d6e3f2c9195bdb943e806442f051d7550aa2578f5c41aa803af Dec 04 04:11:33 crc kubenswrapper[4806]: W1204 04:11:33.234805 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff1a6781_908b_419f_8a02_10abbbfa0816.slice/crio-708f1b71ef9abd789c74e7e56f36d69b7da0bceb2ca6c3b2c12626b8852da10e WatchSource:0}: Error finding container 708f1b71ef9abd789c74e7e56f36d69b7da0bceb2ca6c3b2c12626b8852da10e: Status 404 returned error can't find the container with id 708f1b71ef9abd789c74e7e56f36d69b7da0bceb2ca6c3b2c12626b8852da10e Dec 04 04:11:33 crc kubenswrapper[4806]: I1204 04:11:33.609915 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" event={"ID":"2a7907e6-c36f-4534-9460-80c0091af286","Type":"ContainerStarted","Data":"1bbd41e6a7277b347a6abd9d81d51cb0c06b0fd0907773bb5da72b67b3890e85"} Dec 04 04:11:33 crc kubenswrapper[4806]: I1204 04:11:33.612484 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" event={"ID":"04d2c949-c17a-4299-a9be-b56a7c34866c","Type":"ContainerStarted","Data":"17c01f21b0522d6e3f2c9195bdb943e806442f051d7550aa2578f5c41aa803af"} Dec 04 04:11:33 crc kubenswrapper[4806]: I1204 04:11:33.614680 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" event={"ID":"ff1a6781-908b-419f-8a02-10abbbfa0816","Type":"ContainerStarted","Data":"708f1b71ef9abd789c74e7e56f36d69b7da0bceb2ca6c3b2c12626b8852da10e"} Dec 04 04:11:34 crc kubenswrapper[4806]: I1204 04:11:34.624188 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g2dfw" event={"ID":"b2bba37e-d570-432f-9520-9b5de2005e5c","Type":"ContainerStarted","Data":"9f64ed6815c0dddce4eea86b19061a433d4428e65adcca80ede066c9b026e1d7"} Dec 04 04:11:34 crc kubenswrapper[4806]: I1204 04:11:34.626433 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-69l55" event={"ID":"2473e4f1-90f0-43a0-9e72-0c761d335d76","Type":"ContainerStarted","Data":"1d785a32a294333a9d62425ac6ef847932a9427caf3b59da125a6c5b867f1774"} Dec 04 04:11:34 crc kubenswrapper[4806]: I1204 04:11:34.630229 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vhxjt" event={"ID":"b496f26d-2493-416f-a0f8-5b33e41dbda0","Type":"ContainerStarted","Data":"b1935be2fa35ff46094be948441752a7a9b63f2b9d3e4b27215f4f045fe97faf"} Dec 04 04:11:34 crc kubenswrapper[4806]: I1204 04:11:34.634716 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vtcg5" event={"ID":"084c526b-94d8-43ae-9ffb-33aa215de8c2","Type":"ContainerStarted","Data":"5e45327c7763b0a02e6b0c28b5eebbe75ba2cd5f421997f6a196a4d4ce1d7ac0"} Dec 04 04:11:34 crc kubenswrapper[4806]: I1204 04:11:34.638523 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sktq6" event={"ID":"e06aa87c-9f7d-44f7-b96d-27141d2fff93","Type":"ContainerStarted","Data":"a9d95b51c8e696e88def7e0a01aae7a64bcc7828a769c2241818796f74a1aae3"} Dec 04 04:11:40 crc kubenswrapper[4806]: I1204 04:11:40.923552 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx" event={"ID":"3580e441-1df2-4fe9-8969-e1621fc554dc","Type":"ContainerStarted","Data":"1c16729796acc4eb862016e2171c6236e05e1e23f14d587b495e4f9ebcfdb386"} Dec 04 04:11:40 crc kubenswrapper[4806]: I1204 04:11:40.926345 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" event={"ID":"ff1a6781-908b-419f-8a02-10abbbfa0816","Type":"ContainerStarted","Data":"232da91458d406cae510a0a675a9a4e1da4c0775501d247d3a53e157f4c441cf"} Dec 04 04:11:40 crc kubenswrapper[4806]: I1204 04:11:40.926541 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:11:40 crc kubenswrapper[4806]: I1204 04:11:40.956593 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" podStartSLOduration=41.956575511 podStartE2EDuration="41.956575511s" podCreationTimestamp="2025-12-04 04:10:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:11:40.949093937 +0000 UTC m=+1015.807606865" watchObservedRunningTime="2025-12-04 04:11:40.956575511 +0000 UTC m=+1015.815088439" Dec 04 04:11:42 crc kubenswrapper[4806]: E1204 04:11:42.699515 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 04 04:11:42 crc kubenswrapper[4806]: E1204 04:11:42.699732 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jffl2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-2spbg_openstack-operators(66428f64-e34e-4b37-b911-13e7142f1c18): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:11:42 crc kubenswrapper[4806]: E1204 04:11:42.701089 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2spbg" podUID="66428f64-e34e-4b37-b911-13e7142f1c18" Dec 04 04:11:42 crc kubenswrapper[4806]: E1204 04:11:42.708812 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 04 04:11:42 crc kubenswrapper[4806]: E1204 04:11:42.708983 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wzlmc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-mxmf2_openstack-operators(91ab6e01-027e-4e7a-a250-374b4e8873fc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:11:42 crc kubenswrapper[4806]: E1204 04:11:42.710289 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mxmf2" podUID="91ab6e01-027e-4e7a-a250-374b4e8873fc" Dec 04 04:11:44 crc kubenswrapper[4806]: E1204 04:11:44.120106 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-45jzr" podUID="310d58eb-6345-4cba-8a01-373c65c7bc40" Dec 04 04:11:44 crc kubenswrapper[4806]: E1204 04:11:44.191311 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-25nxx" podUID="a022d699-4a2b-49e2-8d8f-787a3d888df8" Dec 04 04:11:44 crc kubenswrapper[4806]: E1204 04:11:44.204377 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-f6k5f" podUID="5f55ddcf-ca5d-4249-9384-e541ae7d4536" Dec 04 04:11:44 crc kubenswrapper[4806]: E1204 04:11:44.300224 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-xd6pj" podUID="7213e0bc-21af-4df7-929d-f915e343f184" Dec 04 04:11:44 crc kubenswrapper[4806]: E1204 04:11:44.444381 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-kvnjv" podUID="647621de-6ecf-4ebe-96f2-e85f26e4917e" Dec 04 04:11:44 crc kubenswrapper[4806]: E1204 04:11:44.475710 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2lk62" podUID="a3827706-d53c-48f3-b4cf-1c240c0f868c" Dec 04 04:11:44 crc kubenswrapper[4806]: E1204 04:11:44.634208 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kcmvz" podUID="1cfe7f94-e830-4145-a3e7-8a3f694adda1" Dec 04 04:11:44 crc kubenswrapper[4806]: E1204 04:11:44.838136 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2b2qk" podUID="95ae1fa1-3ac7-4040-86da-150daca97f6b" Dec 04 04:11:44 crc kubenswrapper[4806]: I1204 04:11:44.996696 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kcmvz" event={"ID":"1cfe7f94-e830-4145-a3e7-8a3f694adda1","Type":"ContainerStarted","Data":"18a917156ce097744a074ea30ab1ec2f2c74050bf2b631102888a307b7a2e243"} Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.018703 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vhxjt" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.022030 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-45jzr" event={"ID":"310d58eb-6345-4cba-8a01-373c65c7bc40","Type":"ContainerStarted","Data":"a376d56b8087efae418a0a85b6f50ef3fc952705b91babed0277b67640a30c5d"} Dec 04 04:11:45 crc kubenswrapper[4806]: E1204 04:11:45.028466 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-pk7ws" podUID="5f479a93-bb3a-45b8-aa74-60e9f0a7aebd" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.029367 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-xd6pj" event={"ID":"7213e0bc-21af-4df7-929d-f915e343f184","Type":"ContainerStarted","Data":"d5560b17e26fd9eb0f8d7e3aafdd3103bb3fcbce4e61d1875542ab74c76782b2"} Dec 04 04:11:45 crc kubenswrapper[4806]: E1204 04:11:45.031944 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mnhlt" podUID="45192fbb-caf2-4df8-82a6-c222f73b033c" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.038766 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2b2qk" event={"ID":"95ae1fa1-3ac7-4040-86da-150daca97f6b","Type":"ContainerStarted","Data":"e8648996edc276b3c6057bac21f85b941fa92975e01bff042819f66806cec3de"} Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.051779 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2lk62" event={"ID":"a3827706-d53c-48f3-b4cf-1c240c0f868c","Type":"ContainerStarted","Data":"777b122c21dcf9dd6c3f0e669d1d019fd2128363f138883e95e2976a42fdf661"} Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.052795 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vhxjt" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.057838 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-f6k5f" event={"ID":"5f55ddcf-ca5d-4249-9384-e541ae7d4536","Type":"ContainerStarted","Data":"2b0f829462615acfed72a87102e2b365bdaf9b4ee135f087617ae7c71b0f5869"} Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.076208 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2spbg" event={"ID":"66428f64-e34e-4b37-b911-13e7142f1c18","Type":"ContainerStarted","Data":"6eb2f6bdc6b96e61ea119e5ef3b7efdda65ea920de2c19c5f7ef3a1a08d91e3c"} Dec 04 04:11:45 crc kubenswrapper[4806]: E1204 04:11:45.090286 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nt5vq" podUID="669dd56a-69b0-4315-aacc-c8f085f0b76a" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.117358 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vhxjt" podStartSLOduration=4.51931865 podStartE2EDuration="47.117339015s" podCreationTimestamp="2025-12-04 04:10:58 +0000 UTC" firstStartedPulling="2025-12-04 04:11:01.078386296 +0000 UTC m=+975.936899224" lastFinishedPulling="2025-12-04 04:11:43.676406661 +0000 UTC m=+1018.534919589" observedRunningTime="2025-12-04 04:11:45.109553649 +0000 UTC m=+1019.968066577" watchObservedRunningTime="2025-12-04 04:11:45.117339015 +0000 UTC m=+1019.975851943" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.144185 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vtcg5" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.144319 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vtcg5" event={"ID":"084c526b-94d8-43ae-9ffb-33aa215de8c2","Type":"ContainerStarted","Data":"d5c260283848c43d238681c73b35abc754d9f302af5b35927bf71f008d1c5eed"} Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.147888 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vtcg5" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.150495 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mxmf2" event={"ID":"91ab6e01-027e-4e7a-a250-374b4e8873fc","Type":"ContainerStarted","Data":"f733c87419cbe8affb7559fcd6a95cc29615e4140a2eb0ae8f61bb09f4360f50"} Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.151119 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mxmf2" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.156917 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" event={"ID":"2a7907e6-c36f-4534-9460-80c0091af286","Type":"ContainerStarted","Data":"6d07c80f9dd1af4f026efe91a414a7141801e682fd9c30457306351deeb2ad4b"} Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.181045 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" event={"ID":"04d2c949-c17a-4299-a9be-b56a7c34866c","Type":"ContainerStarted","Data":"f23a9a4ab5cda1644f5d4f38fb9bfc3f383cacf4ec478a0d6cb674e1d24730ff"} Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.206231 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx" event={"ID":"3580e441-1df2-4fe9-8969-e1621fc554dc","Type":"ContainerStarted","Data":"83ad8b0ebb5e8a3e2db2e62ebb94c0569cf45c55bb097bf3782a69069006cdaa"} Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.207054 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.218097 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-25nxx" event={"ID":"a022d699-4a2b-49e2-8d8f-787a3d888df8","Type":"ContainerStarted","Data":"cb45b32fe5105795c883b458b4fc192698a1e0b274cb9363b91d9c6084519c31"} Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.218162 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.231256 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-69l55" event={"ID":"2473e4f1-90f0-43a0-9e72-0c761d335d76","Type":"ContainerStarted","Data":"0ab6ec9417cc3ae7bf3f3d5cc60f0b73b5944986248511edba16ae6f724d3f38"} Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.233157 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-69l55" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.236050 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-69l55" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.258070 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-kvnjv" event={"ID":"647621de-6ecf-4ebe-96f2-e85f26e4917e","Type":"ContainerStarted","Data":"b623fb42e5274ff8b262316409ed36319f2a7a420afd85996a438ca1d0b03492"} Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.380352 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-vtcg5" podStartSLOduration=4.730710315 podStartE2EDuration="47.380337648s" podCreationTimestamp="2025-12-04 04:10:58 +0000 UTC" firstStartedPulling="2025-12-04 04:11:00.937317978 +0000 UTC m=+975.795830906" lastFinishedPulling="2025-12-04 04:11:43.586945311 +0000 UTC m=+1018.445458239" observedRunningTime="2025-12-04 04:11:45.336215872 +0000 UTC m=+1020.194728800" watchObservedRunningTime="2025-12-04 04:11:45.380337648 +0000 UTC m=+1020.238850576" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.380903 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-69l55" podStartSLOduration=4.935521394 podStartE2EDuration="46.380896887s" podCreationTimestamp="2025-12-04 04:10:59 +0000 UTC" firstStartedPulling="2025-12-04 04:11:02.141775904 +0000 UTC m=+977.000288832" lastFinishedPulling="2025-12-04 04:11:43.587151397 +0000 UTC m=+1018.445664325" observedRunningTime="2025-12-04 04:11:45.376369838 +0000 UTC m=+1020.234882766" watchObservedRunningTime="2025-12-04 04:11:45.380896887 +0000 UTC m=+1020.239409815" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.444961 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-fdqbx" podStartSLOduration=4.994614649 podStartE2EDuration="46.444943774s" podCreationTimestamp="2025-12-04 04:10:59 +0000 UTC" firstStartedPulling="2025-12-04 04:11:02.160728986 +0000 UTC m=+977.019241914" lastFinishedPulling="2025-12-04 04:11:43.611058111 +0000 UTC m=+1018.469571039" observedRunningTime="2025-12-04 04:11:45.44115335 +0000 UTC m=+1020.299666278" watchObservedRunningTime="2025-12-04 04:11:45.444943774 +0000 UTC m=+1020.303456702" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.565813 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mxmf2" podStartSLOduration=4.914514235 podStartE2EDuration="47.565796562s" podCreationTimestamp="2025-12-04 04:10:58 +0000 UTC" firstStartedPulling="2025-12-04 04:11:00.936724459 +0000 UTC m=+975.795237387" lastFinishedPulling="2025-12-04 04:11:43.588006746 +0000 UTC m=+1018.446519714" observedRunningTime="2025-12-04 04:11:45.563091073 +0000 UTC m=+1020.421604001" watchObservedRunningTime="2025-12-04 04:11:45.565796562 +0000 UTC m=+1020.424309480" Dec 04 04:11:45 crc kubenswrapper[4806]: I1204 04:11:45.746410 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-8d698776b-npcz2" Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.299282 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sktq6" event={"ID":"e06aa87c-9f7d-44f7-b96d-27141d2fff93","Type":"ContainerStarted","Data":"9bb2788fdb4b9f7a834b268f448dbec154d1348ab22395a8b58d64072784205e"} Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.302311 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sktq6" Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.304276 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g2dfw" event={"ID":"b2bba37e-d570-432f-9520-9b5de2005e5c","Type":"ContainerStarted","Data":"18d6687fe258ec744099e40c5ecf334e3dec4b662e458761398c288f2d11c5b0"} Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.304725 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g2dfw" Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.305519 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sktq6" Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.308371 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g2dfw" Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.309022 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2spbg" event={"ID":"66428f64-e34e-4b37-b911-13e7142f1c18","Type":"ContainerStarted","Data":"8ad8743525245fb246eeace4da8afbc0d6a9aabd04c3b6aa173950c516264f06"} Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.309144 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2spbg" Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.310771 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-vhxjt" event={"ID":"b496f26d-2493-416f-a0f8-5b33e41dbda0","Type":"ContainerStarted","Data":"036fa8bfaa40d31a2274f7fc7e0478dc7dbc98c6afb76351539f74ea1f7de33e"} Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.317289 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" event={"ID":"2a7907e6-c36f-4534-9460-80c0091af286","Type":"ContainerStarted","Data":"2eeebcdd098dbd2dcaa8486dc70df55e710f18de301f31549eb1425ff82adeae"} Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.317552 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.319758 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-pk7ws" event={"ID":"5f479a93-bb3a-45b8-aa74-60e9f0a7aebd","Type":"ContainerStarted","Data":"b5bf594cd88b7b7b218a8612a5db8fc2cf7ae921747060235b66078cb11f31fa"} Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.322347 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nt5vq" event={"ID":"669dd56a-69b0-4315-aacc-c8f085f0b76a","Type":"ContainerStarted","Data":"ff7698f8157187cd59a2ad8d709db616a47efbb50670b08212f8a7cab9cf9536"} Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.326675 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-sktq6" podStartSLOduration=6.877432463 podStartE2EDuration="48.326662822s" podCreationTimestamp="2025-12-04 04:10:58 +0000 UTC" firstStartedPulling="2025-12-04 04:11:02.159831526 +0000 UTC m=+977.018344454" lastFinishedPulling="2025-12-04 04:11:43.609061895 +0000 UTC m=+1018.467574813" observedRunningTime="2025-12-04 04:11:46.326072403 +0000 UTC m=+1021.184585341" watchObservedRunningTime="2025-12-04 04:11:46.326662822 +0000 UTC m=+1021.185175750" Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.326835 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" event={"ID":"04d2c949-c17a-4299-a9be-b56a7c34866c","Type":"ContainerStarted","Data":"15e8f1cafed36f7100efbe5ddaedb9ad9cc08bbd52ea7349ef8f21a0b53f5b30"} Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.328046 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.329566 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mnhlt" event={"ID":"45192fbb-caf2-4df8-82a6-c222f73b033c","Type":"ContainerStarted","Data":"8264c232f7d56bdb76303d3607cb2e7a3d53b142e467e90bb76b6c8223951014"} Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.332113 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mxmf2" event={"ID":"91ab6e01-027e-4e7a-a250-374b4e8873fc","Type":"ContainerStarted","Data":"d34fab69ae95e1a1aecb9a2b618cf95ba022b6daf13f76ef0f6709472007df0e"} Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.359058 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2spbg" podStartSLOduration=6.695763143 podStartE2EDuration="48.359028572s" podCreationTimestamp="2025-12-04 04:10:58 +0000 UTC" firstStartedPulling="2025-12-04 04:11:01.922669459 +0000 UTC m=+976.781182387" lastFinishedPulling="2025-12-04 04:11:43.585934888 +0000 UTC m=+1018.444447816" observedRunningTime="2025-12-04 04:11:46.352342054 +0000 UTC m=+1021.210854982" watchObservedRunningTime="2025-12-04 04:11:46.359028572 +0000 UTC m=+1021.217541500" Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.412399 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" podStartSLOduration=38.313619365 podStartE2EDuration="48.412379639s" podCreationTimestamp="2025-12-04 04:10:58 +0000 UTC" firstStartedPulling="2025-12-04 04:11:33.263054404 +0000 UTC m=+1008.121567332" lastFinishedPulling="2025-12-04 04:11:43.361814678 +0000 UTC m=+1018.220327606" observedRunningTime="2025-12-04 04:11:46.407366875 +0000 UTC m=+1021.265879803" watchObservedRunningTime="2025-12-04 04:11:46.412379639 +0000 UTC m=+1021.270892567" Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.490450 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-g2dfw" podStartSLOduration=4.994492453 podStartE2EDuration="48.490431586s" podCreationTimestamp="2025-12-04 04:10:58 +0000 UTC" firstStartedPulling="2025-12-04 04:11:00.113301818 +0000 UTC m=+974.971814746" lastFinishedPulling="2025-12-04 04:11:43.609240951 +0000 UTC m=+1018.467753879" observedRunningTime="2025-12-04 04:11:46.487987856 +0000 UTC m=+1021.346500784" watchObservedRunningTime="2025-12-04 04:11:46.490431586 +0000 UTC m=+1021.348944514" Dec 04 04:11:46 crc kubenswrapper[4806]: I1204 04:11:46.571234 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" podStartSLOduration=38.486350792 podStartE2EDuration="48.571215671s" podCreationTimestamp="2025-12-04 04:10:58 +0000 UTC" firstStartedPulling="2025-12-04 04:11:33.242290023 +0000 UTC m=+1008.100802951" lastFinishedPulling="2025-12-04 04:11:43.327154892 +0000 UTC m=+1018.185667830" observedRunningTime="2025-12-04 04:11:46.566526688 +0000 UTC m=+1021.425039616" watchObservedRunningTime="2025-12-04 04:11:46.571215671 +0000 UTC m=+1021.429728599" Dec 04 04:11:47 crc kubenswrapper[4806]: I1204 04:11:47.344637 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-f6k5f" event={"ID":"5f55ddcf-ca5d-4249-9384-e541ae7d4536","Type":"ContainerStarted","Data":"00aa7355f5143e7c7e3ab1ae787738b3100f1fb76cb0ca999a5608dc8f6cb205"} Dec 04 04:11:47 crc kubenswrapper[4806]: I1204 04:11:47.350081 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2b2qk" event={"ID":"95ae1fa1-3ac7-4040-86da-150daca97f6b","Type":"ContainerStarted","Data":"08eb1678419ade4edbd98d9134569abf140f019de98fb405103cfb073ed2a2a9"} Dec 04 04:11:47 crc kubenswrapper[4806]: I1204 04:11:47.350266 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2b2qk" Dec 04 04:11:47 crc kubenswrapper[4806]: I1204 04:11:47.386393 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2b2qk" podStartSLOduration=4.604167179 podStartE2EDuration="49.386369919s" podCreationTimestamp="2025-12-04 04:10:58 +0000 UTC" firstStartedPulling="2025-12-04 04:11:02.149953363 +0000 UTC m=+977.008466281" lastFinishedPulling="2025-12-04 04:11:46.932156093 +0000 UTC m=+1021.790669021" observedRunningTime="2025-12-04 04:11:47.379732393 +0000 UTC m=+1022.238245321" watchObservedRunningTime="2025-12-04 04:11:47.386369919 +0000 UTC m=+1022.244882847" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.359491 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-25nxx" event={"ID":"a022d699-4a2b-49e2-8d8f-787a3d888df8","Type":"ContainerStarted","Data":"6c15bd88ffca9aaad291c8c8849559d42780d73e8f9be5678d2efaadbda0ee92"} Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.359793 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-25nxx" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.361835 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2lk62" event={"ID":"a3827706-d53c-48f3-b4cf-1c240c0f868c","Type":"ContainerStarted","Data":"9c704ffe02c7172eb4e159f02f28ee5b4d85e029047c3aff1ad26e9bac29fd7a"} Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.362043 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2lk62" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.364016 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-pk7ws" event={"ID":"5f479a93-bb3a-45b8-aa74-60e9f0a7aebd","Type":"ContainerStarted","Data":"617527b10f1b8332ebbfd9506a77554ca788a3ec873ef520c3f0ee9ea436ba62"} Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.364385 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-pk7ws" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.365973 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nt5vq" event={"ID":"669dd56a-69b0-4315-aacc-c8f085f0b76a","Type":"ContainerStarted","Data":"d1d4ed0246deb95559960f81e0e3dba3536de0e90b7a82c6a76501ae43b6c044"} Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.366114 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nt5vq" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.368028 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cwnth" event={"ID":"be179b75-6e57-4537-b3ee-a57dc049948e","Type":"ContainerStarted","Data":"b1f5af91923c4ac0b235a6de3e6bfdcf293ef83c1f8ba019536e7d185bba0726"} Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.370348 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-45jzr" event={"ID":"310d58eb-6345-4cba-8a01-373c65c7bc40","Type":"ContainerStarted","Data":"a936e0913ff78fc8a160656f795c56b1361bbcd21f01a34dfb24a6297143d7f5"} Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.371312 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-45jzr" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.373669 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-xd6pj" event={"ID":"7213e0bc-21af-4df7-929d-f915e343f184","Type":"ContainerStarted","Data":"701e9db6fba7dd40f43dcb27fae8683a3958a2c03b0ae9919bdd0ce326a683bc"} Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.373770 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-xd6pj" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.375411 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-kvnjv" event={"ID":"647621de-6ecf-4ebe-96f2-e85f26e4917e","Type":"ContainerStarted","Data":"6b271afeee2af34d3441207ed5daf8e8497cbb1e606cf15b55277ad5c35a850c"} Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.375842 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-kvnjv" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.379730 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kcmvz" event={"ID":"1cfe7f94-e830-4145-a3e7-8a3f694adda1","Type":"ContainerStarted","Data":"b27400a281ab998beebd7b1596f7d55b1f3d061a41d1cf613e4a0e4f8ec84a4a"} Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.380725 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kcmvz" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.382400 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-25nxx" podStartSLOduration=3.294299118 podStartE2EDuration="50.38238432s" podCreationTimestamp="2025-12-04 04:10:58 +0000 UTC" firstStartedPulling="2025-12-04 04:11:00.361044443 +0000 UTC m=+975.219557371" lastFinishedPulling="2025-12-04 04:11:47.449129645 +0000 UTC m=+1022.307642573" observedRunningTime="2025-12-04 04:11:48.377966426 +0000 UTC m=+1023.236479354" watchObservedRunningTime="2025-12-04 04:11:48.38238432 +0000 UTC m=+1023.240897238" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.383512 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mnhlt" event={"ID":"45192fbb-caf2-4df8-82a6-c222f73b033c","Type":"ContainerStarted","Data":"df45f330ce2b6258bec242e5e1cb363a4407eb53d73feb6628879cfbe30da7c3"} Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.383553 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-f6k5f" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.383568 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mnhlt" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.404808 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-45jzr" podStartSLOduration=5.066076667 podStartE2EDuration="50.404785594s" podCreationTimestamp="2025-12-04 04:10:58 +0000 UTC" firstStartedPulling="2025-12-04 04:11:01.594666796 +0000 UTC m=+976.453179724" lastFinishedPulling="2025-12-04 04:11:46.933375723 +0000 UTC m=+1021.791888651" observedRunningTime="2025-12-04 04:11:48.403350708 +0000 UTC m=+1023.261863646" watchObservedRunningTime="2025-12-04 04:11:48.404785594 +0000 UTC m=+1023.263298532" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.450351 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2lk62" podStartSLOduration=4.66156163 podStartE2EDuration="49.450329496s" podCreationTimestamp="2025-12-04 04:10:59 +0000 UTC" firstStartedPulling="2025-12-04 04:11:02.144911997 +0000 UTC m=+977.003424925" lastFinishedPulling="2025-12-04 04:11:46.933679863 +0000 UTC m=+1021.792192791" observedRunningTime="2025-12-04 04:11:48.435258483 +0000 UTC m=+1023.293771421" watchObservedRunningTime="2025-12-04 04:11:48.450329496 +0000 UTC m=+1023.308842424" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.488201 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nt5vq" podStartSLOduration=5.108157935 podStartE2EDuration="50.488179026s" podCreationTimestamp="2025-12-04 04:10:58 +0000 UTC" firstStartedPulling="2025-12-04 04:11:02.066375115 +0000 UTC m=+976.924888053" lastFinishedPulling="2025-12-04 04:11:47.446396216 +0000 UTC m=+1022.304909144" observedRunningTime="2025-12-04 04:11:48.48646824 +0000 UTC m=+1023.344981178" watchObservedRunningTime="2025-12-04 04:11:48.488179026 +0000 UTC m=+1023.346691954" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.603449 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-pk7ws" podStartSLOduration=5.297870039 podStartE2EDuration="50.60342792s" podCreationTimestamp="2025-12-04 04:10:58 +0000 UTC" firstStartedPulling="2025-12-04 04:11:02.143544623 +0000 UTC m=+977.002057551" lastFinishedPulling="2025-12-04 04:11:47.449102504 +0000 UTC m=+1022.307615432" observedRunningTime="2025-12-04 04:11:48.524819686 +0000 UTC m=+1023.383332624" watchObservedRunningTime="2025-12-04 04:11:48.60342792 +0000 UTC m=+1023.461940848" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.640185 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cwnth" podStartSLOduration=4.362979311 podStartE2EDuration="49.640170984s" podCreationTimestamp="2025-12-04 04:10:59 +0000 UTC" firstStartedPulling="2025-12-04 04:11:01.986058834 +0000 UTC m=+976.844571762" lastFinishedPulling="2025-12-04 04:11:47.263250507 +0000 UTC m=+1022.121763435" observedRunningTime="2025-12-04 04:11:48.637503026 +0000 UTC m=+1023.496015954" watchObservedRunningTime="2025-12-04 04:11:48.640170984 +0000 UTC m=+1023.498683912" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.640312 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-xd6pj" podStartSLOduration=3.757596092 podStartE2EDuration="50.640306438s" podCreationTimestamp="2025-12-04 04:10:58 +0000 UTC" firstStartedPulling="2025-12-04 04:11:00.04954178 +0000 UTC m=+974.908054708" lastFinishedPulling="2025-12-04 04:11:46.932252126 +0000 UTC m=+1021.790765054" observedRunningTime="2025-12-04 04:11:48.606237892 +0000 UTC m=+1023.464750820" watchObservedRunningTime="2025-12-04 04:11:48.640306438 +0000 UTC m=+1023.498819366" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.684483 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-kvnjv" podStartSLOduration=4.264233698 podStartE2EDuration="49.684461225s" podCreationTimestamp="2025-12-04 04:10:59 +0000 UTC" firstStartedPulling="2025-12-04 04:11:02.039858897 +0000 UTC m=+976.898371825" lastFinishedPulling="2025-12-04 04:11:47.460086424 +0000 UTC m=+1022.318599352" observedRunningTime="2025-12-04 04:11:48.679288575 +0000 UTC m=+1023.537801513" watchObservedRunningTime="2025-12-04 04:11:48.684461225 +0000 UTC m=+1023.542974163" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.729028 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-f6k5f" podStartSLOduration=4.763433474 podStartE2EDuration="50.729007594s" podCreationTimestamp="2025-12-04 04:10:58 +0000 UTC" firstStartedPulling="2025-12-04 04:11:00.967825954 +0000 UTC m=+975.826338882" lastFinishedPulling="2025-12-04 04:11:46.933400074 +0000 UTC m=+1021.791913002" observedRunningTime="2025-12-04 04:11:48.727462303 +0000 UTC m=+1023.585975231" watchObservedRunningTime="2025-12-04 04:11:48.729007594 +0000 UTC m=+1023.587520522" Dec 04 04:11:48 crc kubenswrapper[4806]: I1204 04:11:48.831825 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mnhlt" podStartSLOduration=4.164403607 podStartE2EDuration="49.83180499s" podCreationTimestamp="2025-12-04 04:10:59 +0000 UTC" firstStartedPulling="2025-12-04 04:11:01.781708102 +0000 UTC m=+976.640221030" lastFinishedPulling="2025-12-04 04:11:47.449109485 +0000 UTC m=+1022.307622413" observedRunningTime="2025-12-04 04:11:48.771776024 +0000 UTC m=+1023.630288952" watchObservedRunningTime="2025-12-04 04:11:48.83180499 +0000 UTC m=+1023.690317918" Dec 04 04:11:49 crc kubenswrapper[4806]: I1204 04:11:49.140639 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mxmf2" Dec 04 04:11:49 crc kubenswrapper[4806]: I1204 04:11:49.170096 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kcmvz" podStartSLOduration=6.029341576 podStartE2EDuration="51.170076669s" podCreationTimestamp="2025-12-04 04:10:58 +0000 UTC" firstStartedPulling="2025-12-04 04:11:01.793818149 +0000 UTC m=+976.652331077" lastFinishedPulling="2025-12-04 04:11:46.934553252 +0000 UTC m=+1021.793066170" observedRunningTime="2025-12-04 04:11:48.832423581 +0000 UTC m=+1023.690936499" watchObservedRunningTime="2025-12-04 04:11:49.170076669 +0000 UTC m=+1024.028589597" Dec 04 04:11:49 crc kubenswrapper[4806]: I1204 04:11:49.672483 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2spbg" Dec 04 04:11:54 crc kubenswrapper[4806]: I1204 04:11:54.938482 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-g6svq" Dec 04 04:11:55 crc kubenswrapper[4806]: I1204 04:11:55.343123 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7" Dec 04 04:11:58 crc kubenswrapper[4806]: I1204 04:11:58.845799 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-xd6pj" Dec 04 04:11:59 crc kubenswrapper[4806]: I1204 04:11:59.064483 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-25nxx" Dec 04 04:11:59 crc kubenswrapper[4806]: I1204 04:11:59.156389 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-f6k5f" Dec 04 04:11:59 crc kubenswrapper[4806]: I1204 04:11:59.491808 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-45jzr" Dec 04 04:11:59 crc kubenswrapper[4806]: I1204 04:11:59.512089 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-2b2qk" Dec 04 04:11:59 crc kubenswrapper[4806]: I1204 04:11:59.530007 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-pk7ws" Dec 04 04:11:59 crc kubenswrapper[4806]: I1204 04:11:59.663289 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-kcmvz" Dec 04 04:11:59 crc kubenswrapper[4806]: I1204 04:11:59.691588 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nt5vq" Dec 04 04:11:59 crc kubenswrapper[4806]: I1204 04:11:59.973609 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2lk62" Dec 04 04:11:59 crc kubenswrapper[4806]: I1204 04:11:59.982977 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-kvnjv" Dec 04 04:12:00 crc kubenswrapper[4806]: I1204 04:12:00.090752 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-mnhlt" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.606886 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4k495"] Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.609050 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4k495" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.618028 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.618183 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.618497 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.619004 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-8jfdc" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.645220 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4k495"] Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.747612 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-4d8g8"] Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.747989 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0851aee-3856-4b79-8e25-5285bcb05470-config\") pod \"dnsmasq-dns-675f4bcbfc-4k495\" (UID: \"f0851aee-3856-4b79-8e25-5285bcb05470\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4k495" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.748043 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kxk8\" (UniqueName: \"kubernetes.io/projected/f0851aee-3856-4b79-8e25-5285bcb05470-kube-api-access-6kxk8\") pod \"dnsmasq-dns-675f4bcbfc-4k495\" (UID: \"f0851aee-3856-4b79-8e25-5285bcb05470\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4k495" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.749073 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-4d8g8" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.753248 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.777687 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-4d8g8"] Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.849439 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94snc\" (UniqueName: \"kubernetes.io/projected/88599a5d-02e0-4c25-85e3-26d886cad1a4-kube-api-access-94snc\") pod \"dnsmasq-dns-78dd6ddcc-4d8g8\" (UID: \"88599a5d-02e0-4c25-85e3-26d886cad1a4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-4d8g8" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.849527 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88599a5d-02e0-4c25-85e3-26d886cad1a4-config\") pod \"dnsmasq-dns-78dd6ddcc-4d8g8\" (UID: \"88599a5d-02e0-4c25-85e3-26d886cad1a4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-4d8g8" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.849692 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0851aee-3856-4b79-8e25-5285bcb05470-config\") pod \"dnsmasq-dns-675f4bcbfc-4k495\" (UID: \"f0851aee-3856-4b79-8e25-5285bcb05470\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4k495" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.849779 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kxk8\" (UniqueName: \"kubernetes.io/projected/f0851aee-3856-4b79-8e25-5285bcb05470-kube-api-access-6kxk8\") pod \"dnsmasq-dns-675f4bcbfc-4k495\" (UID: \"f0851aee-3856-4b79-8e25-5285bcb05470\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4k495" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.849878 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88599a5d-02e0-4c25-85e3-26d886cad1a4-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-4d8g8\" (UID: \"88599a5d-02e0-4c25-85e3-26d886cad1a4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-4d8g8" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.850778 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0851aee-3856-4b79-8e25-5285bcb05470-config\") pod \"dnsmasq-dns-675f4bcbfc-4k495\" (UID: \"f0851aee-3856-4b79-8e25-5285bcb05470\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4k495" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.886522 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kxk8\" (UniqueName: \"kubernetes.io/projected/f0851aee-3856-4b79-8e25-5285bcb05470-kube-api-access-6kxk8\") pod \"dnsmasq-dns-675f4bcbfc-4k495\" (UID: \"f0851aee-3856-4b79-8e25-5285bcb05470\") " pod="openstack/dnsmasq-dns-675f4bcbfc-4k495" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.940461 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4k495" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.951805 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94snc\" (UniqueName: \"kubernetes.io/projected/88599a5d-02e0-4c25-85e3-26d886cad1a4-kube-api-access-94snc\") pod \"dnsmasq-dns-78dd6ddcc-4d8g8\" (UID: \"88599a5d-02e0-4c25-85e3-26d886cad1a4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-4d8g8" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.951871 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88599a5d-02e0-4c25-85e3-26d886cad1a4-config\") pod \"dnsmasq-dns-78dd6ddcc-4d8g8\" (UID: \"88599a5d-02e0-4c25-85e3-26d886cad1a4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-4d8g8" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.951991 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88599a5d-02e0-4c25-85e3-26d886cad1a4-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-4d8g8\" (UID: \"88599a5d-02e0-4c25-85e3-26d886cad1a4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-4d8g8" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.953383 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88599a5d-02e0-4c25-85e3-26d886cad1a4-config\") pod \"dnsmasq-dns-78dd6ddcc-4d8g8\" (UID: \"88599a5d-02e0-4c25-85e3-26d886cad1a4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-4d8g8" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.953982 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88599a5d-02e0-4c25-85e3-26d886cad1a4-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-4d8g8\" (UID: \"88599a5d-02e0-4c25-85e3-26d886cad1a4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-4d8g8" Dec 04 04:12:15 crc kubenswrapper[4806]: I1204 04:12:15.974881 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94snc\" (UniqueName: \"kubernetes.io/projected/88599a5d-02e0-4c25-85e3-26d886cad1a4-kube-api-access-94snc\") pod \"dnsmasq-dns-78dd6ddcc-4d8g8\" (UID: \"88599a5d-02e0-4c25-85e3-26d886cad1a4\") " pod="openstack/dnsmasq-dns-78dd6ddcc-4d8g8" Dec 04 04:12:16 crc kubenswrapper[4806]: I1204 04:12:16.064732 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-4d8g8" Dec 04 04:12:16 crc kubenswrapper[4806]: I1204 04:12:16.525400 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4k495"] Dec 04 04:12:16 crc kubenswrapper[4806]: I1204 04:12:16.578022 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-4k495" event={"ID":"f0851aee-3856-4b79-8e25-5285bcb05470","Type":"ContainerStarted","Data":"92a1750c0536c97e74457d5b4db25aa78b463a4cc9d2034082e106640c87a594"} Dec 04 04:12:16 crc kubenswrapper[4806]: I1204 04:12:16.646585 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-4d8g8"] Dec 04 04:12:16 crc kubenswrapper[4806]: W1204 04:12:16.650063 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88599a5d_02e0_4c25_85e3_26d886cad1a4.slice/crio-ad55e6b3851ac3dee44a98c134385d0b6bd8f933c527c86de35e259170eb31cb WatchSource:0}: Error finding container ad55e6b3851ac3dee44a98c134385d0b6bd8f933c527c86de35e259170eb31cb: Status 404 returned error can't find the container with id ad55e6b3851ac3dee44a98c134385d0b6bd8f933c527c86de35e259170eb31cb Dec 04 04:12:17 crc kubenswrapper[4806]: I1204 04:12:17.634400 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-4d8g8" event={"ID":"88599a5d-02e0-4c25-85e3-26d886cad1a4","Type":"ContainerStarted","Data":"ad55e6b3851ac3dee44a98c134385d0b6bd8f933c527c86de35e259170eb31cb"} Dec 04 04:12:17 crc kubenswrapper[4806]: I1204 04:12:17.733938 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4k495"] Dec 04 04:12:17 crc kubenswrapper[4806]: I1204 04:12:17.757030 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-ht488"] Dec 04 04:12:17 crc kubenswrapper[4806]: I1204 04:12:17.758316 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" Dec 04 04:12:17 crc kubenswrapper[4806]: I1204 04:12:17.770892 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-ht488"] Dec 04 04:12:17 crc kubenswrapper[4806]: I1204 04:12:17.788993 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-ht488\" (UID: \"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b\") " pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" Dec 04 04:12:17 crc kubenswrapper[4806]: I1204 04:12:17.789088 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-config\") pod \"dnsmasq-dns-5ccc8479f9-ht488\" (UID: \"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b\") " pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" Dec 04 04:12:17 crc kubenswrapper[4806]: I1204 04:12:17.789109 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jq79\" (UniqueName: \"kubernetes.io/projected/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-kube-api-access-6jq79\") pod \"dnsmasq-dns-5ccc8479f9-ht488\" (UID: \"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b\") " pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" Dec 04 04:12:17 crc kubenswrapper[4806]: I1204 04:12:17.890327 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-config\") pod \"dnsmasq-dns-5ccc8479f9-ht488\" (UID: \"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b\") " pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" Dec 04 04:12:17 crc kubenswrapper[4806]: I1204 04:12:17.890372 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jq79\" (UniqueName: \"kubernetes.io/projected/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-kube-api-access-6jq79\") pod \"dnsmasq-dns-5ccc8479f9-ht488\" (UID: \"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b\") " pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" Dec 04 04:12:17 crc kubenswrapper[4806]: I1204 04:12:17.890434 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-ht488\" (UID: \"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b\") " pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" Dec 04 04:12:17 crc kubenswrapper[4806]: I1204 04:12:17.891260 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-ht488\" (UID: \"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b\") " pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" Dec 04 04:12:17 crc kubenswrapper[4806]: I1204 04:12:17.891777 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-config\") pod \"dnsmasq-dns-5ccc8479f9-ht488\" (UID: \"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b\") " pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" Dec 04 04:12:17 crc kubenswrapper[4806]: I1204 04:12:17.918235 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jq79\" (UniqueName: \"kubernetes.io/projected/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-kube-api-access-6jq79\") pod \"dnsmasq-dns-5ccc8479f9-ht488\" (UID: \"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b\") " pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.126669 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.130794 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-4d8g8"] Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.151889 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z2hxt"] Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.155746 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.175465 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z2hxt"] Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.207583 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcg2z\" (UniqueName: \"kubernetes.io/projected/3b288f07-5df3-4373-bb91-631b4cf409a4-kube-api-access-hcg2z\") pod \"dnsmasq-dns-57d769cc4f-z2hxt\" (UID: \"3b288f07-5df3-4373-bb91-631b4cf409a4\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.207630 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b288f07-5df3-4373-bb91-631b4cf409a4-config\") pod \"dnsmasq-dns-57d769cc4f-z2hxt\" (UID: \"3b288f07-5df3-4373-bb91-631b4cf409a4\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.207671 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b288f07-5df3-4373-bb91-631b4cf409a4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-z2hxt\" (UID: \"3b288f07-5df3-4373-bb91-631b4cf409a4\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.308522 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcg2z\" (UniqueName: \"kubernetes.io/projected/3b288f07-5df3-4373-bb91-631b4cf409a4-kube-api-access-hcg2z\") pod \"dnsmasq-dns-57d769cc4f-z2hxt\" (UID: \"3b288f07-5df3-4373-bb91-631b4cf409a4\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.308589 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b288f07-5df3-4373-bb91-631b4cf409a4-config\") pod \"dnsmasq-dns-57d769cc4f-z2hxt\" (UID: \"3b288f07-5df3-4373-bb91-631b4cf409a4\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.308621 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b288f07-5df3-4373-bb91-631b4cf409a4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-z2hxt\" (UID: \"3b288f07-5df3-4373-bb91-631b4cf409a4\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.309512 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b288f07-5df3-4373-bb91-631b4cf409a4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-z2hxt\" (UID: \"3b288f07-5df3-4373-bb91-631b4cf409a4\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.312163 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b288f07-5df3-4373-bb91-631b4cf409a4-config\") pod \"dnsmasq-dns-57d769cc4f-z2hxt\" (UID: \"3b288f07-5df3-4373-bb91-631b4cf409a4\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.344265 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcg2z\" (UniqueName: \"kubernetes.io/projected/3b288f07-5df3-4373-bb91-631b4cf409a4-kube-api-access-hcg2z\") pod \"dnsmasq-dns-57d769cc4f-z2hxt\" (UID: \"3b288f07-5df3-4373-bb91-631b4cf409a4\") " pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.548128 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.829509 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-ht488"] Dec 04 04:12:18 crc kubenswrapper[4806]: W1204 04:12:18.847870 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1e0b1f6_f4a8_4791_8b68_bf16dfa00c8b.slice/crio-d195b2fb206399fb071c5cc8aeb1b830f25e80fbee8a3cc0133a5de92ece4c0c WatchSource:0}: Error finding container d195b2fb206399fb071c5cc8aeb1b830f25e80fbee8a3cc0133a5de92ece4c0c: Status 404 returned error can't find the container with id d195b2fb206399fb071c5cc8aeb1b830f25e80fbee8a3cc0133a5de92ece4c0c Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.938578 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.942846 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.960441 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.963762 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.963852 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.965520 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.967370 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.967663 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.968227 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-52qcl" Dec 04 04:12:18 crc kubenswrapper[4806]: I1204 04:12:18.970033 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.028800 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25bnr\" (UniqueName: \"kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-kube-api-access-25bnr\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.028882 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.028966 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.029912 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f18b8381-c1f9-4e48-9639-adfc5a406726-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.030078 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.030114 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.030157 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.030208 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.030234 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f18b8381-c1f9-4e48-9639-adfc5a406726-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.030311 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.030351 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.123284 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z2hxt"] Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.132897 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f18b8381-c1f9-4e48-9639-adfc5a406726-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.133085 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.133119 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.133150 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.133175 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f18b8381-c1f9-4e48-9639-adfc5a406726-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.133203 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.133797 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.133886 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.133991 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25bnr\" (UniqueName: \"kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-kube-api-access-25bnr\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.134085 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.134170 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.134402 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.135304 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.135334 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.135573 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.135790 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.136707 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.142373 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f18b8381-c1f9-4e48-9639-adfc5a406726-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.144182 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.146599 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f18b8381-c1f9-4e48-9639-adfc5a406726-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.147231 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: W1204 04:12:19.147632 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b288f07_5df3_4373_bb91_631b4cf409a4.slice/crio-9940b55480ee4ec450a86e3e810c42a5024c27b828eb3940b040437a9579a988 WatchSource:0}: Error finding container 9940b55480ee4ec450a86e3e810c42a5024c27b828eb3940b040437a9579a988: Status 404 returned error can't find the container with id 9940b55480ee4ec450a86e3e810c42a5024c27b828eb3940b040437a9579a988 Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.165116 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25bnr\" (UniqueName: \"kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-kube-api-access-25bnr\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.176601 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.279658 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.332564 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.333923 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.342113 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.342282 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.345796 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.345870 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.346177 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.350818 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-xqlhc" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.351743 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.369138 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.448676 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.448724 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.448754 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.448780 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.448801 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7bc6\" (UniqueName: \"kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-kube-api-access-l7bc6\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.448816 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-config-data\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.448833 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.448853 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.449081 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.449132 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.449158 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.550414 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.550489 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.550510 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.550537 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.550565 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.550595 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7bc6\" (UniqueName: \"kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-kube-api-access-l7bc6\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.550615 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-config-data\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.550635 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.550654 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.550713 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.550764 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.553023 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.555896 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.556562 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-config-data\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.558119 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.568865 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.569284 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.573521 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.581305 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.585291 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.588555 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.588878 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7bc6\" (UniqueName: \"kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-kube-api-access-l7bc6\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.633272 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.680249 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.837115 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" event={"ID":"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b","Type":"ContainerStarted","Data":"d195b2fb206399fb071c5cc8aeb1b830f25e80fbee8a3cc0133a5de92ece4c0c"} Dec 04 04:12:19 crc kubenswrapper[4806]: I1204 04:12:19.904817 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" event={"ID":"3b288f07-5df3-4373-bb91-631b4cf409a4","Type":"ContainerStarted","Data":"9940b55480ee4ec450a86e3e810c42a5024c27b828eb3940b040437a9579a988"} Dec 04 04:12:20 crc kubenswrapper[4806]: I1204 04:12:20.425959 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 04:12:20 crc kubenswrapper[4806]: I1204 04:12:20.598980 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 04:12:20 crc kubenswrapper[4806]: W1204 04:12:20.600955 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b7d7449_e532_4a32_bfcd_dc9f92767bd6.slice/crio-00831f055f546ac605c4cbaecc3d90db3164d4e16e66117fa45ee873406b097d WatchSource:0}: Error finding container 00831f055f546ac605c4cbaecc3d90db3164d4e16e66117fa45ee873406b097d: Status 404 returned error can't find the container with id 00831f055f546ac605c4cbaecc3d90db3164d4e16e66117fa45ee873406b097d Dec 04 04:12:20 crc kubenswrapper[4806]: I1204 04:12:20.842186 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 04 04:12:20 crc kubenswrapper[4806]: I1204 04:12:20.846096 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 04:12:20 crc kubenswrapper[4806]: I1204 04:12:20.857819 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-b242h" Dec 04 04:12:20 crc kubenswrapper[4806]: I1204 04:12:20.857971 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 04 04:12:20 crc kubenswrapper[4806]: I1204 04:12:20.858095 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 04 04:12:20 crc kubenswrapper[4806]: I1204 04:12:20.858356 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 04 04:12:20 crc kubenswrapper[4806]: I1204 04:12:20.861092 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 04 04:12:20 crc kubenswrapper[4806]: I1204 04:12:20.870983 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 04:12:20 crc kubenswrapper[4806]: I1204 04:12:20.940570 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0b7d7449-e532-4a32-bfcd-dc9f92767bd6","Type":"ContainerStarted","Data":"00831f055f546ac605c4cbaecc3d90db3164d4e16e66117fa45ee873406b097d"} Dec 04 04:12:20 crc kubenswrapper[4806]: I1204 04:12:20.959509 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f18b8381-c1f9-4e48-9639-adfc5a406726","Type":"ContainerStarted","Data":"b820a3f39cc9a745c86cb6837feaa651ad78664699cb08440440e659544397fa"} Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.008333 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.008446 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-config-data-default\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.008476 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.008510 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.008560 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.008581 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nrhs\" (UniqueName: \"kubernetes.io/projected/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-kube-api-access-4nrhs\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.008608 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.008639 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-kolla-config\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.110118 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-config-data-default\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.110191 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.110244 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.110310 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.110354 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nrhs\" (UniqueName: \"kubernetes.io/projected/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-kube-api-access-4nrhs\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.110391 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.110430 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-kolla-config\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.110500 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.112592 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.116752 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.117112 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.117827 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-kolla-config\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.129132 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-config-data-default\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.145029 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.145871 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nrhs\" (UniqueName: \"kubernetes.io/projected/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-kube-api-access-4nrhs\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.162080 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99326fd5-69d5-4e7d-b5c6-f7808344ad6d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.167343 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"99326fd5-69d5-4e7d-b5c6-f7808344ad6d\") " pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.186779 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.975137 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.980577 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.988513 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.988708 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 04 04:12:21 crc kubenswrapper[4806]: I1204 04:12:21.988843 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.001134 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-8zcv8" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.022608 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.047816 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.098520 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.099466 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.103060 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.103214 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.103255 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-ssfx9" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.134003 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d1a2b83-04d7-4db8-acd1-456cc0c33206-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.134058 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0d1a2b83-04d7-4db8-acd1-456cc0c33206-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.134090 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d1a2b83-04d7-4db8-acd1-456cc0c33206-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.134112 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwl2c\" (UniqueName: \"kubernetes.io/projected/0d1a2b83-04d7-4db8-acd1-456cc0c33206-kube-api-access-xwl2c\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.134145 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0d1a2b83-04d7-4db8-acd1-456cc0c33206-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.134162 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0d1a2b83-04d7-4db8-acd1-456cc0c33206-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.134201 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d1a2b83-04d7-4db8-acd1-456cc0c33206-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.134223 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.138905 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.235185 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0face013-e59b-4e32-ace5-d64d45302e7c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0face013-e59b-4e32-ace5-d64d45302e7c\") " pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.235233 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d1a2b83-04d7-4db8-acd1-456cc0c33206-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.235259 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0face013-e59b-4e32-ace5-d64d45302e7c-config-data\") pod \"memcached-0\" (UID: \"0face013-e59b-4e32-ace5-d64d45302e7c\") " pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.235279 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0face013-e59b-4e32-ace5-d64d45302e7c-kolla-config\") pod \"memcached-0\" (UID: \"0face013-e59b-4e32-ace5-d64d45302e7c\") " pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.235309 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0d1a2b83-04d7-4db8-acd1-456cc0c33206-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.235368 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d1a2b83-04d7-4db8-acd1-456cc0c33206-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.235387 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwl2c\" (UniqueName: \"kubernetes.io/projected/0d1a2b83-04d7-4db8-acd1-456cc0c33206-kube-api-access-xwl2c\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.235437 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grxsf\" (UniqueName: \"kubernetes.io/projected/0face013-e59b-4e32-ace5-d64d45302e7c-kube-api-access-grxsf\") pod \"memcached-0\" (UID: \"0face013-e59b-4e32-ace5-d64d45302e7c\") " pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.235463 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0d1a2b83-04d7-4db8-acd1-456cc0c33206-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.235481 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0d1a2b83-04d7-4db8-acd1-456cc0c33206-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.235523 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d1a2b83-04d7-4db8-acd1-456cc0c33206-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.235545 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.235568 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0face013-e59b-4e32-ace5-d64d45302e7c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0face013-e59b-4e32-ace5-d64d45302e7c\") " pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.236346 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0d1a2b83-04d7-4db8-acd1-456cc0c33206-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.236808 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0d1a2b83-04d7-4db8-acd1-456cc0c33206-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.237037 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0d1a2b83-04d7-4db8-acd1-456cc0c33206-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.237054 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.237510 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d1a2b83-04d7-4db8-acd1-456cc0c33206-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.250530 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d1a2b83-04d7-4db8-acd1-456cc0c33206-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.275071 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d1a2b83-04d7-4db8-acd1-456cc0c33206-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.297418 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwl2c\" (UniqueName: \"kubernetes.io/projected/0d1a2b83-04d7-4db8-acd1-456cc0c33206-kube-api-access-xwl2c\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.341506 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grxsf\" (UniqueName: \"kubernetes.io/projected/0face013-e59b-4e32-ace5-d64d45302e7c-kube-api-access-grxsf\") pod \"memcached-0\" (UID: \"0face013-e59b-4e32-ace5-d64d45302e7c\") " pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.341617 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0face013-e59b-4e32-ace5-d64d45302e7c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0face013-e59b-4e32-ace5-d64d45302e7c\") " pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.341652 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0face013-e59b-4e32-ace5-d64d45302e7c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0face013-e59b-4e32-ace5-d64d45302e7c\") " pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.341674 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0face013-e59b-4e32-ace5-d64d45302e7c-config-data\") pod \"memcached-0\" (UID: \"0face013-e59b-4e32-ace5-d64d45302e7c\") " pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.341692 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0face013-e59b-4e32-ace5-d64d45302e7c-kolla-config\") pod \"memcached-0\" (UID: \"0face013-e59b-4e32-ace5-d64d45302e7c\") " pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.342465 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0face013-e59b-4e32-ace5-d64d45302e7c-kolla-config\") pod \"memcached-0\" (UID: \"0face013-e59b-4e32-ace5-d64d45302e7c\") " pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.347325 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0face013-e59b-4e32-ace5-d64d45302e7c-config-data\") pod \"memcached-0\" (UID: \"0face013-e59b-4e32-ace5-d64d45302e7c\") " pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.386344 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grxsf\" (UniqueName: \"kubernetes.io/projected/0face013-e59b-4e32-ace5-d64d45302e7c-kube-api-access-grxsf\") pod \"memcached-0\" (UID: \"0face013-e59b-4e32-ace5-d64d45302e7c\") " pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.387715 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"0d1a2b83-04d7-4db8-acd1-456cc0c33206\") " pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.397720 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0face013-e59b-4e32-ace5-d64d45302e7c-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0face013-e59b-4e32-ace5-d64d45302e7c\") " pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.398181 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0face013-e59b-4e32-ace5-d64d45302e7c-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0face013-e59b-4e32-ace5-d64d45302e7c\") " pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.473461 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 04:12:22 crc kubenswrapper[4806]: I1204 04:12:22.633409 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 04:12:23 crc kubenswrapper[4806]: I1204 04:12:23.053535 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"99326fd5-69d5-4e7d-b5c6-f7808344ad6d","Type":"ContainerStarted","Data":"dc19bd9d0c8479c5efc945fd6cb0cff02a490829bf5f2a8159507844be5198d3"} Dec 04 04:12:23 crc kubenswrapper[4806]: I1204 04:12:23.104915 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 04 04:12:23 crc kubenswrapper[4806]: W1204 04:12:23.267153 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0face013_e59b_4e32_ace5_d64d45302e7c.slice/crio-9b71a204ec6e28e41aa936ee8bac28a8bfa392602f6a128e5a2dfd99d15b0c3a WatchSource:0}: Error finding container 9b71a204ec6e28e41aa936ee8bac28a8bfa392602f6a128e5a2dfd99d15b0c3a: Status 404 returned error can't find the container with id 9b71a204ec6e28e41aa936ee8bac28a8bfa392602f6a128e5a2dfd99d15b0c3a Dec 04 04:12:23 crc kubenswrapper[4806]: I1204 04:12:23.404032 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 04:12:24 crc kubenswrapper[4806]: I1204 04:12:24.022805 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 04:12:24 crc kubenswrapper[4806]: I1204 04:12:24.025362 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 04:12:24 crc kubenswrapper[4806]: I1204 04:12:24.040437 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 04:12:24 crc kubenswrapper[4806]: I1204 04:12:24.040717 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-4vg4m" Dec 04 04:12:24 crc kubenswrapper[4806]: I1204 04:12:24.098725 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlljs\" (UniqueName: \"kubernetes.io/projected/fe13c4d3-85c5-44aa-8424-415f4a775165-kube-api-access-tlljs\") pod \"kube-state-metrics-0\" (UID: \"fe13c4d3-85c5-44aa-8424-415f4a775165\") " pod="openstack/kube-state-metrics-0" Dec 04 04:12:24 crc kubenswrapper[4806]: I1204 04:12:24.156645 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0d1a2b83-04d7-4db8-acd1-456cc0c33206","Type":"ContainerStarted","Data":"7f7c38dcc0eeeb6467ca23d49612650df86588e23cc2d1fb9dd7d5d846f647b3"} Dec 04 04:12:24 crc kubenswrapper[4806]: I1204 04:12:24.162705 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0face013-e59b-4e32-ace5-d64d45302e7c","Type":"ContainerStarted","Data":"9b71a204ec6e28e41aa936ee8bac28a8bfa392602f6a128e5a2dfd99d15b0c3a"} Dec 04 04:12:24 crc kubenswrapper[4806]: I1204 04:12:24.200635 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlljs\" (UniqueName: \"kubernetes.io/projected/fe13c4d3-85c5-44aa-8424-415f4a775165-kube-api-access-tlljs\") pod \"kube-state-metrics-0\" (UID: \"fe13c4d3-85c5-44aa-8424-415f4a775165\") " pod="openstack/kube-state-metrics-0" Dec 04 04:12:24 crc kubenswrapper[4806]: I1204 04:12:24.262108 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlljs\" (UniqueName: \"kubernetes.io/projected/fe13c4d3-85c5-44aa-8424-415f4a775165-kube-api-access-tlljs\") pod \"kube-state-metrics-0\" (UID: \"fe13c4d3-85c5-44aa-8424-415f4a775165\") " pod="openstack/kube-state-metrics-0" Dec 04 04:12:24 crc kubenswrapper[4806]: I1204 04:12:24.412859 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 04:12:25 crc kubenswrapper[4806]: I1204 04:12:25.124993 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 04:12:25 crc kubenswrapper[4806]: W1204 04:12:25.157361 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe13c4d3_85c5_44aa_8424_415f4a775165.slice/crio-619d60dd9550135fd596223448d353f678ea95b7d4dbe4f31b4b7561c63cfd2f WatchSource:0}: Error finding container 619d60dd9550135fd596223448d353f678ea95b7d4dbe4f31b4b7561c63cfd2f: Status 404 returned error can't find the container with id 619d60dd9550135fd596223448d353f678ea95b7d4dbe4f31b4b7561c63cfd2f Dec 04 04:12:25 crc kubenswrapper[4806]: I1204 04:12:25.200316 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fe13c4d3-85c5-44aa-8424-415f4a775165","Type":"ContainerStarted","Data":"619d60dd9550135fd596223448d353f678ea95b7d4dbe4f31b4b7561c63cfd2f"} Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.047700 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.048442 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.535185 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.538639 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.541831 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.576148 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.576266 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-68kqp" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.576443 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.588802 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.591049 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.721445 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75e0b5e3-e994-4ef3-9673-539ca96485c5-config\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.721536 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75e0b5e3-e994-4ef3-9673-539ca96485c5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.721568 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/75e0b5e3-e994-4ef3-9673-539ca96485c5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.721592 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75e0b5e3-e994-4ef3-9673-539ca96485c5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.721631 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/75e0b5e3-e994-4ef3-9673-539ca96485c5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.721655 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49dzk\" (UniqueName: \"kubernetes.io/projected/75e0b5e3-e994-4ef3-9673-539ca96485c5-kube-api-access-49dzk\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.721773 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/75e0b5e3-e994-4ef3-9673-539ca96485c5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.721857 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.823398 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75e0b5e3-e994-4ef3-9673-539ca96485c5-config\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.823515 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75e0b5e3-e994-4ef3-9673-539ca96485c5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.823548 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/75e0b5e3-e994-4ef3-9673-539ca96485c5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.823609 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75e0b5e3-e994-4ef3-9673-539ca96485c5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.823647 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/75e0b5e3-e994-4ef3-9673-539ca96485c5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.823692 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49dzk\" (UniqueName: \"kubernetes.io/projected/75e0b5e3-e994-4ef3-9673-539ca96485c5-kube-api-access-49dzk\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.823727 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/75e0b5e3-e994-4ef3-9673-539ca96485c5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.823767 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.824146 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.825536 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/75e0b5e3-e994-4ef3-9673-539ca96485c5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.825702 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75e0b5e3-e994-4ef3-9673-539ca96485c5-config\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.827123 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75e0b5e3-e994-4ef3-9673-539ca96485c5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.833822 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75e0b5e3-e994-4ef3-9673-539ca96485c5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.834115 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/75e0b5e3-e994-4ef3-9673-539ca96485c5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.837541 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/75e0b5e3-e994-4ef3-9673-539ca96485c5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.848983 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49dzk\" (UniqueName: \"kubernetes.io/projected/75e0b5e3-e994-4ef3-9673-539ca96485c5-kube-api-access-49dzk\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.867386 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"75e0b5e3-e994-4ef3-9673-539ca96485c5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:27 crc kubenswrapper[4806]: I1204 04:12:27.914791 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.772495 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tx487"] Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.776120 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.784138 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-bt2n7" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.790077 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.790350 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.821064 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tx487"] Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.827628 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-2sz9t"] Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.829480 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.851954 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ef210587-097b-48d8-8097-dc1b95e2456e-var-run-ovn\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.852490 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ef210587-097b-48d8-8097-dc1b95e2456e-var-run\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.852636 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef210587-097b-48d8-8097-dc1b95e2456e-combined-ca-bundle\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.852707 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf9z5\" (UniqueName: \"kubernetes.io/projected/ef210587-097b-48d8-8097-dc1b95e2456e-kube-api-access-jf9z5\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.852795 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef210587-097b-48d8-8097-dc1b95e2456e-ovn-controller-tls-certs\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.852853 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef210587-097b-48d8-8097-dc1b95e2456e-scripts\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.852896 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ef210587-097b-48d8-8097-dc1b95e2456e-var-log-ovn\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.869813 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-2sz9t"] Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.956486 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/63e4ca21-d6b3-4e93-965e-3dde5e35027f-etc-ovs\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.956565 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/63e4ca21-d6b3-4e93-965e-3dde5e35027f-var-lib\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.956616 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ef210587-097b-48d8-8097-dc1b95e2456e-var-run-ovn\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.956661 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ef210587-097b-48d8-8097-dc1b95e2456e-var-run\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.956730 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef210587-097b-48d8-8097-dc1b95e2456e-combined-ca-bundle\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.956759 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf9z5\" (UniqueName: \"kubernetes.io/projected/ef210587-097b-48d8-8097-dc1b95e2456e-kube-api-access-jf9z5\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.956794 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63e4ca21-d6b3-4e93-965e-3dde5e35027f-scripts\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.956834 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/63e4ca21-d6b3-4e93-965e-3dde5e35027f-var-log\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.956880 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef210587-097b-48d8-8097-dc1b95e2456e-ovn-controller-tls-certs\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.956917 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef210587-097b-48d8-8097-dc1b95e2456e-scripts\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.956967 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/63e4ca21-d6b3-4e93-965e-3dde5e35027f-var-run\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.957003 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ef210587-097b-48d8-8097-dc1b95e2456e-var-log-ovn\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.964062 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ef210587-097b-48d8-8097-dc1b95e2456e-var-log-ovn\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.964325 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ef210587-097b-48d8-8097-dc1b95e2456e-var-run-ovn\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.964914 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ef210587-097b-48d8-8097-dc1b95e2456e-var-run\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.966123 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78kfb\" (UniqueName: \"kubernetes.io/projected/63e4ca21-d6b3-4e93-965e-3dde5e35027f-kube-api-access-78kfb\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.968920 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef210587-097b-48d8-8097-dc1b95e2456e-scripts\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:28 crc kubenswrapper[4806]: I1204 04:12:28.983285 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef210587-097b-48d8-8097-dc1b95e2456e-combined-ca-bundle\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.013355 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef210587-097b-48d8-8097-dc1b95e2456e-ovn-controller-tls-certs\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.034541 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf9z5\" (UniqueName: \"kubernetes.io/projected/ef210587-097b-48d8-8097-dc1b95e2456e-kube-api-access-jf9z5\") pod \"ovn-controller-tx487\" (UID: \"ef210587-097b-48d8-8097-dc1b95e2456e\") " pod="openstack/ovn-controller-tx487" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.068353 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/63e4ca21-d6b3-4e93-965e-3dde5e35027f-var-log\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.068449 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/63e4ca21-d6b3-4e93-965e-3dde5e35027f-var-run\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.068500 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78kfb\" (UniqueName: \"kubernetes.io/projected/63e4ca21-d6b3-4e93-965e-3dde5e35027f-kube-api-access-78kfb\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.068552 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/63e4ca21-d6b3-4e93-965e-3dde5e35027f-etc-ovs\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.068582 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/63e4ca21-d6b3-4e93-965e-3dde5e35027f-var-lib\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.068678 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63e4ca21-d6b3-4e93-965e-3dde5e35027f-scripts\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.069521 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/63e4ca21-d6b3-4e93-965e-3dde5e35027f-var-log\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.069589 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/63e4ca21-d6b3-4e93-965e-3dde5e35027f-var-run\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.069750 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/63e4ca21-d6b3-4e93-965e-3dde5e35027f-var-lib\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.071186 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/63e4ca21-d6b3-4e93-965e-3dde5e35027f-etc-ovs\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.077365 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63e4ca21-d6b3-4e93-965e-3dde5e35027f-scripts\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.106253 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78kfb\" (UniqueName: \"kubernetes.io/projected/63e4ca21-d6b3-4e93-965e-3dde5e35027f-kube-api-access-78kfb\") pod \"ovn-controller-ovs-2sz9t\" (UID: \"63e4ca21-d6b3-4e93-965e-3dde5e35027f\") " pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.122776 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tx487" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.160326 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:29 crc kubenswrapper[4806]: I1204 04:12:29.506393 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 04:12:30 crc kubenswrapper[4806]: W1204 04:12:30.139447 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75e0b5e3_e994_4ef3_9673_539ca96485c5.slice/crio-d07a3dd4f7971a46e71b19ba5a02ed870134a75fc7734a740c2809ef240a347f WatchSource:0}: Error finding container d07a3dd4f7971a46e71b19ba5a02ed870134a75fc7734a740c2809ef240a347f: Status 404 returned error can't find the container with id d07a3dd4f7971a46e71b19ba5a02ed870134a75fc7734a740c2809ef240a347f Dec 04 04:12:30 crc kubenswrapper[4806]: I1204 04:12:30.356774 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"75e0b5e3-e994-4ef3-9673-539ca96485c5","Type":"ContainerStarted","Data":"d07a3dd4f7971a46e71b19ba5a02ed870134a75fc7734a740c2809ef240a347f"} Dec 04 04:12:31 crc kubenswrapper[4806]: I1204 04:12:31.893718 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 04:12:31 crc kubenswrapper[4806]: I1204 04:12:31.896492 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:31 crc kubenswrapper[4806]: I1204 04:12:31.899783 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 04 04:12:31 crc kubenswrapper[4806]: I1204 04:12:31.899894 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-4c6b4" Dec 04 04:12:31 crc kubenswrapper[4806]: I1204 04:12:31.900029 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 04 04:12:31 crc kubenswrapper[4806]: I1204 04:12:31.899992 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 04 04:12:31 crc kubenswrapper[4806]: I1204 04:12:31.901263 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.052149 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.052222 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09739947-6f7d-4a4e-8a85-05a8a7345894-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.052270 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/09739947-6f7d-4a4e-8a85-05a8a7345894-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.052308 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr6gk\" (UniqueName: \"kubernetes.io/projected/09739947-6f7d-4a4e-8a85-05a8a7345894-kube-api-access-fr6gk\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.052435 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09739947-6f7d-4a4e-8a85-05a8a7345894-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.052611 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09739947-6f7d-4a4e-8a85-05a8a7345894-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.052683 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/09739947-6f7d-4a4e-8a85-05a8a7345894-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.053014 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09739947-6f7d-4a4e-8a85-05a8a7345894-config\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.156869 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09739947-6f7d-4a4e-8a85-05a8a7345894-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.159202 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/09739947-6f7d-4a4e-8a85-05a8a7345894-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.159490 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09739947-6f7d-4a4e-8a85-05a8a7345894-config\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.159636 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.159773 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09739947-6f7d-4a4e-8a85-05a8a7345894-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.159899 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/09739947-6f7d-4a4e-8a85-05a8a7345894-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.160056 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr6gk\" (UniqueName: \"kubernetes.io/projected/09739947-6f7d-4a4e-8a85-05a8a7345894-kube-api-access-fr6gk\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.160166 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09739947-6f7d-4a4e-8a85-05a8a7345894-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.161594 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.162107 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09739947-6f7d-4a4e-8a85-05a8a7345894-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.165623 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/09739947-6f7d-4a4e-8a85-05a8a7345894-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.170634 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09739947-6f7d-4a4e-8a85-05a8a7345894-config\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.177478 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09739947-6f7d-4a4e-8a85-05a8a7345894-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.178343 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09739947-6f7d-4a4e-8a85-05a8a7345894-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.195901 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/09739947-6f7d-4a4e-8a85-05a8a7345894-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.201225 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr6gk\" (UniqueName: \"kubernetes.io/projected/09739947-6f7d-4a4e-8a85-05a8a7345894-kube-api-access-fr6gk\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.218194 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"09739947-6f7d-4a4e-8a85-05a8a7345894\") " pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:32 crc kubenswrapper[4806]: I1204 04:12:32.245748 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 04:12:43 crc kubenswrapper[4806]: E1204 04:12:43.899985 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 04 04:12:43 crc kubenswrapper[4806]: E1204 04:12:43.901645 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-25bnr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(f18b8381-c1f9-4e48-9639-adfc5a406726): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:12:43 crc kubenswrapper[4806]: E1204 04:12:43.903075 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="f18b8381-c1f9-4e48-9639-adfc5a406726" Dec 04 04:12:43 crc kubenswrapper[4806]: E1204 04:12:43.920991 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 04 04:12:43 crc kubenswrapper[4806]: E1204 04:12:43.921160 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l7bc6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(0b7d7449-e532-4a32-bfcd-dc9f92767bd6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:12:43 crc kubenswrapper[4806]: E1204 04:12:43.922637 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="0b7d7449-e532-4a32-bfcd-dc9f92767bd6" Dec 04 04:12:43 crc kubenswrapper[4806]: E1204 04:12:43.940744 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 04 04:12:43 crc kubenswrapper[4806]: E1204 04:12:43.940870 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xwl2c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(0d1a2b83-04d7-4db8-acd1-456cc0c33206): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:12:43 crc kubenswrapper[4806]: E1204 04:12:43.942153 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="0d1a2b83-04d7-4db8-acd1-456cc0c33206" Dec 04 04:12:44 crc kubenswrapper[4806]: E1204 04:12:44.488534 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="f18b8381-c1f9-4e48-9639-adfc5a406726" Dec 04 04:12:44 crc kubenswrapper[4806]: E1204 04:12:44.488589 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="0d1a2b83-04d7-4db8-acd1-456cc0c33206" Dec 04 04:12:44 crc kubenswrapper[4806]: E1204 04:12:44.488653 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="0b7d7449-e532-4a32-bfcd-dc9f92767bd6" Dec 04 04:12:44 crc kubenswrapper[4806]: E1204 04:12:44.619684 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Dec 04 04:12:44 crc kubenswrapper[4806]: E1204 04:12:44.619901 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:nc5hdbh5d4h547h697h96h6bh579h595h67fh676h577hf5h5d4h5c7h7fhddh5f7h68h644h99h5fch697h68dh56fh598h5dhch689h667h67dh5d4q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-grxsf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(0face013-e59b-4e32-ace5-d64d45302e7c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:12:44 crc kubenswrapper[4806]: E1204 04:12:44.623079 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="0face013-e59b-4e32-ace5-d64d45302e7c" Dec 04 04:12:45 crc kubenswrapper[4806]: E1204 04:12:45.496885 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="0face013-e59b-4e32-ace5-d64d45302e7c" Dec 04 04:12:51 crc kubenswrapper[4806]: I1204 04:12:51.445766 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tx487"] Dec 04 04:12:51 crc kubenswrapper[4806]: I1204 04:12:51.698587 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-2sz9t"] Dec 04 04:12:52 crc kubenswrapper[4806]: E1204 04:12:52.006876 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 04 04:12:52 crc kubenswrapper[4806]: E1204 04:12:52.007052 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6kxk8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-4k495_openstack(f0851aee-3856-4b79-8e25-5285bcb05470): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:12:52 crc kubenswrapper[4806]: E1204 04:12:52.008295 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-4k495" podUID="f0851aee-3856-4b79-8e25-5285bcb05470" Dec 04 04:12:52 crc kubenswrapper[4806]: E1204 04:12:52.298081 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 04 04:12:52 crc kubenswrapper[4806]: E1204 04:12:52.298626 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hcg2z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-z2hxt_openstack(3b288f07-5df3-4373-bb91-631b4cf409a4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:12:52 crc kubenswrapper[4806]: E1204 04:12:52.299898 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" podUID="3b288f07-5df3-4373-bb91-631b4cf409a4" Dec 04 04:12:52 crc kubenswrapper[4806]: E1204 04:12:52.311297 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 04 04:12:52 crc kubenswrapper[4806]: E1204 04:12:52.311530 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-94snc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-4d8g8_openstack(88599a5d-02e0-4c25-85e3-26d886cad1a4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:12:52 crc kubenswrapper[4806]: E1204 04:12:52.312904 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-4d8g8" podUID="88599a5d-02e0-4c25-85e3-26d886cad1a4" Dec 04 04:12:52 crc kubenswrapper[4806]: W1204 04:12:52.366720 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e4ca21_d6b3_4e93_965e_3dde5e35027f.slice/crio-6275f594f82f92114ca587dac090ff10c136218b61b81c35c98e61f5473113ba WatchSource:0}: Error finding container 6275f594f82f92114ca587dac090ff10c136218b61b81c35c98e61f5473113ba: Status 404 returned error can't find the container with id 6275f594f82f92114ca587dac090ff10c136218b61b81c35c98e61f5473113ba Dec 04 04:12:52 crc kubenswrapper[4806]: E1204 04:12:52.420318 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 04 04:12:52 crc kubenswrapper[4806]: E1204 04:12:52.420517 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6jq79,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-ht488_openstack(c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:12:52 crc kubenswrapper[4806]: E1204 04:12:52.421660 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" podUID="c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b" Dec 04 04:12:52 crc kubenswrapper[4806]: I1204 04:12:52.560232 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2sz9t" event={"ID":"63e4ca21-d6b3-4e93-965e-3dde5e35027f","Type":"ContainerStarted","Data":"6275f594f82f92114ca587dac090ff10c136218b61b81c35c98e61f5473113ba"} Dec 04 04:12:52 crc kubenswrapper[4806]: I1204 04:12:52.563230 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tx487" event={"ID":"ef210587-097b-48d8-8097-dc1b95e2456e","Type":"ContainerStarted","Data":"56a83912cd4a950a438f0a378014cede4e270d5dc50ea8f0cb8839a7988beaef"} Dec 04 04:12:52 crc kubenswrapper[4806]: E1204 04:12:52.564632 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" podUID="c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b" Dec 04 04:12:52 crc kubenswrapper[4806]: E1204 04:12:52.565644 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" podUID="3b288f07-5df3-4373-bb91-631b4cf409a4" Dec 04 04:12:52 crc kubenswrapper[4806]: I1204 04:12:52.783605 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 04:12:53 crc kubenswrapper[4806]: E1204 04:12:53.210059 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Dec 04 04:12:53 crc kubenswrapper[4806]: E1204 04:12:53.210113 4806 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Dec 04 04:12:53 crc kubenswrapper[4806]: E1204 04:12:53.210260 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tlljs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(fe13c4d3-85c5-44aa-8424-415f4a775165): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 04:12:53 crc kubenswrapper[4806]: E1204 04:12:53.211420 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="fe13c4d3-85c5-44aa-8424-415f4a775165" Dec 04 04:12:53 crc kubenswrapper[4806]: W1204 04:12:53.212677 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09739947_6f7d_4a4e_8a85_05a8a7345894.slice/crio-e6a7c08dbac5b2676cfe29a9ea26545dfaad8d69787218f1953170ae7fd659d8 WatchSource:0}: Error finding container e6a7c08dbac5b2676cfe29a9ea26545dfaad8d69787218f1953170ae7fd659d8: Status 404 returned error can't find the container with id e6a7c08dbac5b2676cfe29a9ea26545dfaad8d69787218f1953170ae7fd659d8 Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.298155 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4k495" Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.313219 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-4d8g8" Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.498593 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kxk8\" (UniqueName: \"kubernetes.io/projected/f0851aee-3856-4b79-8e25-5285bcb05470-kube-api-access-6kxk8\") pod \"f0851aee-3856-4b79-8e25-5285bcb05470\" (UID: \"f0851aee-3856-4b79-8e25-5285bcb05470\") " Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.499607 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88599a5d-02e0-4c25-85e3-26d886cad1a4-dns-svc\") pod \"88599a5d-02e0-4c25-85e3-26d886cad1a4\" (UID: \"88599a5d-02e0-4c25-85e3-26d886cad1a4\") " Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.499743 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0851aee-3856-4b79-8e25-5285bcb05470-config\") pod \"f0851aee-3856-4b79-8e25-5285bcb05470\" (UID: \"f0851aee-3856-4b79-8e25-5285bcb05470\") " Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.499851 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88599a5d-02e0-4c25-85e3-26d886cad1a4-config\") pod \"88599a5d-02e0-4c25-85e3-26d886cad1a4\" (UID: \"88599a5d-02e0-4c25-85e3-26d886cad1a4\") " Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.500137 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94snc\" (UniqueName: \"kubernetes.io/projected/88599a5d-02e0-4c25-85e3-26d886cad1a4-kube-api-access-94snc\") pod \"88599a5d-02e0-4c25-85e3-26d886cad1a4\" (UID: \"88599a5d-02e0-4c25-85e3-26d886cad1a4\") " Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.501553 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88599a5d-02e0-4c25-85e3-26d886cad1a4-config" (OuterVolumeSpecName: "config") pod "88599a5d-02e0-4c25-85e3-26d886cad1a4" (UID: "88599a5d-02e0-4c25-85e3-26d886cad1a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.502018 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0851aee-3856-4b79-8e25-5285bcb05470-config" (OuterVolumeSpecName: "config") pod "f0851aee-3856-4b79-8e25-5285bcb05470" (UID: "f0851aee-3856-4b79-8e25-5285bcb05470"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.502806 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88599a5d-02e0-4c25-85e3-26d886cad1a4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "88599a5d-02e0-4c25-85e3-26d886cad1a4" (UID: "88599a5d-02e0-4c25-85e3-26d886cad1a4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.511405 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0851aee-3856-4b79-8e25-5285bcb05470-kube-api-access-6kxk8" (OuterVolumeSpecName: "kube-api-access-6kxk8") pod "f0851aee-3856-4b79-8e25-5285bcb05470" (UID: "f0851aee-3856-4b79-8e25-5285bcb05470"). InnerVolumeSpecName "kube-api-access-6kxk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.511959 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88599a5d-02e0-4c25-85e3-26d886cad1a4-kube-api-access-94snc" (OuterVolumeSpecName: "kube-api-access-94snc") pod "88599a5d-02e0-4c25-85e3-26d886cad1a4" (UID: "88599a5d-02e0-4c25-85e3-26d886cad1a4"). InnerVolumeSpecName "kube-api-access-94snc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.571800 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-4k495" Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.575041 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"09739947-6f7d-4a4e-8a85-05a8a7345894","Type":"ContainerStarted","Data":"e6a7c08dbac5b2676cfe29a9ea26545dfaad8d69787218f1953170ae7fd659d8"} Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.575087 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-4k495" event={"ID":"f0851aee-3856-4b79-8e25-5285bcb05470","Type":"ContainerDied","Data":"92a1750c0536c97e74457d5b4db25aa78b463a4cc9d2034082e106640c87a594"} Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.580277 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-4d8g8" Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.580281 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-4d8g8" event={"ID":"88599a5d-02e0-4c25-85e3-26d886cad1a4","Type":"ContainerDied","Data":"ad55e6b3851ac3dee44a98c134385d0b6bd8f933c527c86de35e259170eb31cb"} Dec 04 04:12:53 crc kubenswrapper[4806]: E1204 04:12:53.584900 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="fe13c4d3-85c5-44aa-8424-415f4a775165" Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.617355 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94snc\" (UniqueName: \"kubernetes.io/projected/88599a5d-02e0-4c25-85e3-26d886cad1a4-kube-api-access-94snc\") on node \"crc\" DevicePath \"\"" Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.617396 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kxk8\" (UniqueName: \"kubernetes.io/projected/f0851aee-3856-4b79-8e25-5285bcb05470-kube-api-access-6kxk8\") on node \"crc\" DevicePath \"\"" Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.617439 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88599a5d-02e0-4c25-85e3-26d886cad1a4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.617453 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0851aee-3856-4b79-8e25-5285bcb05470-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.617466 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88599a5d-02e0-4c25-85e3-26d886cad1a4-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.657233 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4k495"] Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.663355 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-4k495"] Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.724010 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-4d8g8"] Dec 04 04:12:53 crc kubenswrapper[4806]: I1204 04:12:53.727414 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-4d8g8"] Dec 04 04:12:54 crc kubenswrapper[4806]: I1204 04:12:54.591747 4806 generic.go:334] "Generic (PLEG): container finished" podID="63e4ca21-d6b3-4e93-965e-3dde5e35027f" containerID="3312b884cb8ca07863a4f3912b932bdf9c5e736274d84da6cccecab1e6b6e971" exitCode=0 Dec 04 04:12:54 crc kubenswrapper[4806]: I1204 04:12:54.591908 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2sz9t" event={"ID":"63e4ca21-d6b3-4e93-965e-3dde5e35027f","Type":"ContainerDied","Data":"3312b884cb8ca07863a4f3912b932bdf9c5e736274d84da6cccecab1e6b6e971"} Dec 04 04:12:54 crc kubenswrapper[4806]: I1204 04:12:54.595049 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"75e0b5e3-e994-4ef3-9673-539ca96485c5","Type":"ContainerStarted","Data":"f7fec339afc3563a2412d606620f0df3b52d96fd68128dac3ec72e496554b7e5"} Dec 04 04:12:54 crc kubenswrapper[4806]: I1204 04:12:54.597130 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"99326fd5-69d5-4e7d-b5c6-f7808344ad6d","Type":"ContainerStarted","Data":"8c0b091b45c19179f024ac16ab7c92e7eb89e42463810be3348d48ec67bee471"} Dec 04 04:12:55 crc kubenswrapper[4806]: I1204 04:12:55.451168 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88599a5d-02e0-4c25-85e3-26d886cad1a4" path="/var/lib/kubelet/pods/88599a5d-02e0-4c25-85e3-26d886cad1a4/volumes" Dec 04 04:12:55 crc kubenswrapper[4806]: I1204 04:12:55.451895 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0851aee-3856-4b79-8e25-5285bcb05470" path="/var/lib/kubelet/pods/f0851aee-3856-4b79-8e25-5285bcb05470/volumes" Dec 04 04:12:56 crc kubenswrapper[4806]: I1204 04:12:56.621657 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2sz9t" event={"ID":"63e4ca21-d6b3-4e93-965e-3dde5e35027f","Type":"ContainerStarted","Data":"be1314064c77d3a8bf2ec58f32fecfad31d034adf35f298300599d39fca2807e"} Dec 04 04:12:56 crc kubenswrapper[4806]: I1204 04:12:56.622119 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-2sz9t" event={"ID":"63e4ca21-d6b3-4e93-965e-3dde5e35027f","Type":"ContainerStarted","Data":"4affe7178c16204c395bfa6ac7928af91dbf7c969c3a680db36312e010e0be51"} Dec 04 04:12:56 crc kubenswrapper[4806]: I1204 04:12:56.622133 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:56 crc kubenswrapper[4806]: I1204 04:12:56.622143 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:12:56 crc kubenswrapper[4806]: I1204 04:12:56.624894 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tx487" event={"ID":"ef210587-097b-48d8-8097-dc1b95e2456e","Type":"ContainerStarted","Data":"e6c0433af1a2f096a372071cc74ac42591f8fb284a86fcd85825dde8b3480910"} Dec 04 04:12:56 crc kubenswrapper[4806]: I1204 04:12:56.625608 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-tx487" Dec 04 04:12:56 crc kubenswrapper[4806]: I1204 04:12:56.627422 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"09739947-6f7d-4a4e-8a85-05a8a7345894","Type":"ContainerStarted","Data":"085583c55c877b800bb7a961627cfec2eb2f3e49d8f7d211cf221cfecf836ecd"} Dec 04 04:12:56 crc kubenswrapper[4806]: I1204 04:12:56.649158 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-2sz9t" podStartSLOduration=27.687585114 podStartE2EDuration="28.649133395s" podCreationTimestamp="2025-12-04 04:12:28 +0000 UTC" firstStartedPulling="2025-12-04 04:12:52.401002813 +0000 UTC m=+1087.259515741" lastFinishedPulling="2025-12-04 04:12:53.362551094 +0000 UTC m=+1088.221064022" observedRunningTime="2025-12-04 04:12:56.644041874 +0000 UTC m=+1091.502554802" watchObservedRunningTime="2025-12-04 04:12:56.649133395 +0000 UTC m=+1091.507646333" Dec 04 04:12:56 crc kubenswrapper[4806]: I1204 04:12:56.667245 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-tx487" podStartSLOduration=25.645807578 podStartE2EDuration="28.667227136s" podCreationTimestamp="2025-12-04 04:12:28 +0000 UTC" firstStartedPulling="2025-12-04 04:12:52.401984034 +0000 UTC m=+1087.260496962" lastFinishedPulling="2025-12-04 04:12:55.423403592 +0000 UTC m=+1090.281916520" observedRunningTime="2025-12-04 04:12:56.661211376 +0000 UTC m=+1091.519724304" watchObservedRunningTime="2025-12-04 04:12:56.667227136 +0000 UTC m=+1091.525740064" Dec 04 04:12:57 crc kubenswrapper[4806]: I1204 04:12:57.047432 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:12:57 crc kubenswrapper[4806]: I1204 04:12:57.047496 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:12:57 crc kubenswrapper[4806]: I1204 04:12:57.639436 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0b7d7449-e532-4a32-bfcd-dc9f92767bd6","Type":"ContainerStarted","Data":"e59290c926c0f12d51a460b02e72a020b41534db66e0140ba00eddb2776df964"} Dec 04 04:13:00 crc kubenswrapper[4806]: I1204 04:13:00.668388 4806 generic.go:334] "Generic (PLEG): container finished" podID="99326fd5-69d5-4e7d-b5c6-f7808344ad6d" containerID="8c0b091b45c19179f024ac16ab7c92e7eb89e42463810be3348d48ec67bee471" exitCode=0 Dec 04 04:13:00 crc kubenswrapper[4806]: I1204 04:13:00.669032 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"99326fd5-69d5-4e7d-b5c6-f7808344ad6d","Type":"ContainerDied","Data":"8c0b091b45c19179f024ac16ab7c92e7eb89e42463810be3348d48ec67bee471"} Dec 04 04:13:02 crc kubenswrapper[4806]: I1204 04:13:02.689675 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0d1a2b83-04d7-4db8-acd1-456cc0c33206","Type":"ContainerStarted","Data":"37f7b3f064bdd4c7f6ea365eccc649da0af3e90a691266d3ba657007fed82e06"} Dec 04 04:13:02 crc kubenswrapper[4806]: I1204 04:13:02.693121 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"75e0b5e3-e994-4ef3-9673-539ca96485c5","Type":"ContainerStarted","Data":"b3efafc5cfcfabf6ef15fd1026b683aa4ac4a9786264cc696ca06600f2744ba6"} Dec 04 04:13:02 crc kubenswrapper[4806]: I1204 04:13:02.696431 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0face013-e59b-4e32-ace5-d64d45302e7c","Type":"ContainerStarted","Data":"1a81064de6c6e5759e33a36ad4c9e0b194f9329c0d654958f382bc96de01c8c0"} Dec 04 04:13:02 crc kubenswrapper[4806]: I1204 04:13:02.696645 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 04 04:13:02 crc kubenswrapper[4806]: I1204 04:13:02.702176 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"99326fd5-69d5-4e7d-b5c6-f7808344ad6d","Type":"ContainerStarted","Data":"0a1cebd08ce6656291e17346330c451cc9a22361befc3cf900bcaee075cd849e"} Dec 04 04:13:02 crc kubenswrapper[4806]: I1204 04:13:02.706805 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"09739947-6f7d-4a4e-8a85-05a8a7345894","Type":"ContainerStarted","Data":"87e3bcdd79eb9645e03d4382282707f1f759dd92372e3046f47a4a48d76388d1"} Dec 04 04:13:02 crc kubenswrapper[4806]: I1204 04:13:02.751478 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.809506051 podStartE2EDuration="36.751451431s" podCreationTimestamp="2025-12-04 04:12:26 +0000 UTC" firstStartedPulling="2025-12-04 04:12:30.144596218 +0000 UTC m=+1065.003109146" lastFinishedPulling="2025-12-04 04:13:02.086541598 +0000 UTC m=+1096.945054526" observedRunningTime="2025-12-04 04:13:02.747753373 +0000 UTC m=+1097.606266311" watchObservedRunningTime="2025-12-04 04:13:02.751451431 +0000 UTC m=+1097.609964359" Dec 04 04:13:02 crc kubenswrapper[4806]: I1204 04:13:02.786670 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=1.9501097889999999 podStartE2EDuration="40.786642593s" podCreationTimestamp="2025-12-04 04:12:22 +0000 UTC" firstStartedPulling="2025-12-04 04:12:23.232592924 +0000 UTC m=+1058.091105852" lastFinishedPulling="2025-12-04 04:13:02.069125728 +0000 UTC m=+1096.927638656" observedRunningTime="2025-12-04 04:13:02.776691219 +0000 UTC m=+1097.635204157" watchObservedRunningTime="2025-12-04 04:13:02.786642593 +0000 UTC m=+1097.645155521" Dec 04 04:13:02 crc kubenswrapper[4806]: I1204 04:13:02.819681 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=14.964682328 podStartE2EDuration="43.819639307s" podCreationTimestamp="2025-12-04 04:12:19 +0000 UTC" firstStartedPulling="2025-12-04 04:12:22.059748992 +0000 UTC m=+1056.918261920" lastFinishedPulling="2025-12-04 04:12:50.914705971 +0000 UTC m=+1085.773218899" observedRunningTime="2025-12-04 04:13:02.811958163 +0000 UTC m=+1097.670471111" watchObservedRunningTime="2025-12-04 04:13:02.819639307 +0000 UTC m=+1097.678152235" Dec 04 04:13:02 crc kubenswrapper[4806]: I1204 04:13:02.861276 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=24.011532153 podStartE2EDuration="32.861251053s" podCreationTimestamp="2025-12-04 04:12:30 +0000 UTC" firstStartedPulling="2025-12-04 04:12:53.217178738 +0000 UTC m=+1088.075691666" lastFinishedPulling="2025-12-04 04:13:02.066897628 +0000 UTC m=+1096.925410566" observedRunningTime="2025-12-04 04:13:02.856674377 +0000 UTC m=+1097.715187305" watchObservedRunningTime="2025-12-04 04:13:02.861251053 +0000 UTC m=+1097.719763981" Dec 04 04:13:02 crc kubenswrapper[4806]: I1204 04:13:02.915231 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 04 04:13:03 crc kubenswrapper[4806]: I1204 04:13:03.719426 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f18b8381-c1f9-4e48-9639-adfc5a406726","Type":"ContainerStarted","Data":"12c5f4cdfff6da63c40c019002d93c0627dc153a33cc9d065f4ac4c3b969019d"} Dec 04 04:13:03 crc kubenswrapper[4806]: I1204 04:13:03.915751 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 04 04:13:03 crc kubenswrapper[4806]: I1204 04:13:03.959687 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 04 04:13:04 crc kubenswrapper[4806]: I1204 04:13:04.769158 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.252744 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.314203 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-ht488"] Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.354100 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-xfnvf"] Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.356083 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.358779 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.384002 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-xfnvf"] Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.394819 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.414538 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-4kl52"] Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.418960 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.436819 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.474390 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-4kl52"] Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.542605 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-xfnvf\" (UID: \"711393c3-c7a7-4beb-9971-dea2c573379a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.542662 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-config\") pod \"dnsmasq-dns-7fd796d7df-xfnvf\" (UID: \"711393c3-c7a7-4beb-9971-dea2c573379a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.542718 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9xfp\" (UniqueName: \"kubernetes.io/projected/711393c3-c7a7-4beb-9971-dea2c573379a-kube-api-access-r9xfp\") pod \"dnsmasq-dns-7fd796d7df-xfnvf\" (UID: \"711393c3-c7a7-4beb-9971-dea2c573379a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.542738 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-xfnvf\" (UID: \"711393c3-c7a7-4beb-9971-dea2c573379a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.644325 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0b91c83f-0036-42a8-97b4-86846f14466a-ovs-rundir\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.644952 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-xfnvf\" (UID: \"711393c3-c7a7-4beb-9971-dea2c573379a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.644985 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-config\") pod \"dnsmasq-dns-7fd796d7df-xfnvf\" (UID: \"711393c3-c7a7-4beb-9971-dea2c573379a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.645037 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6w4b\" (UniqueName: \"kubernetes.io/projected/0b91c83f-0036-42a8-97b4-86846f14466a-kube-api-access-g6w4b\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.645106 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b91c83f-0036-42a8-97b4-86846f14466a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.645147 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9xfp\" (UniqueName: \"kubernetes.io/projected/711393c3-c7a7-4beb-9971-dea2c573379a-kube-api-access-r9xfp\") pod \"dnsmasq-dns-7fd796d7df-xfnvf\" (UID: \"711393c3-c7a7-4beb-9971-dea2c573379a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.645179 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-xfnvf\" (UID: \"711393c3-c7a7-4beb-9971-dea2c573379a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.645211 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b91c83f-0036-42a8-97b4-86846f14466a-config\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.645310 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b91c83f-0036-42a8-97b4-86846f14466a-combined-ca-bundle\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.645389 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0b91c83f-0036-42a8-97b4-86846f14466a-ovn-rundir\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.646301 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-xfnvf\" (UID: \"711393c3-c7a7-4beb-9971-dea2c573379a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.647703 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-config\") pod \"dnsmasq-dns-7fd796d7df-xfnvf\" (UID: \"711393c3-c7a7-4beb-9971-dea2c573379a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.650577 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-xfnvf\" (UID: \"711393c3-c7a7-4beb-9971-dea2c573379a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.681041 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9xfp\" (UniqueName: \"kubernetes.io/projected/711393c3-c7a7-4beb-9971-dea2c573379a-kube-api-access-r9xfp\") pod \"dnsmasq-dns-7fd796d7df-xfnvf\" (UID: \"711393c3-c7a7-4beb-9971-dea2c573379a\") " pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.747646 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6w4b\" (UniqueName: \"kubernetes.io/projected/0b91c83f-0036-42a8-97b4-86846f14466a-kube-api-access-g6w4b\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.747782 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b91c83f-0036-42a8-97b4-86846f14466a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.747848 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b91c83f-0036-42a8-97b4-86846f14466a-config\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.748156 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b91c83f-0036-42a8-97b4-86846f14466a-combined-ca-bundle\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.748219 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0b91c83f-0036-42a8-97b4-86846f14466a-ovn-rundir\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.748303 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0b91c83f-0036-42a8-97b4-86846f14466a-ovs-rundir\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.749145 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0b91c83f-0036-42a8-97b4-86846f14466a-ovs-rundir\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.750910 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0b91c83f-0036-42a8-97b4-86846f14466a-ovn-rundir\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.751546 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b91c83f-0036-42a8-97b4-86846f14466a-config\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.756407 4806 generic.go:334] "Generic (PLEG): container finished" podID="3b288f07-5df3-4373-bb91-631b4cf409a4" containerID="a5d4327af30e54de906adbc592216521a6ac03d0058bbab62b67ae65d3c87e8f" exitCode=0 Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.756539 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" event={"ID":"3b288f07-5df3-4373-bb91-631b4cf409a4","Type":"ContainerDied","Data":"a5d4327af30e54de906adbc592216521a6ac03d0058bbab62b67ae65d3c87e8f"} Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.758477 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b91c83f-0036-42a8-97b4-86846f14466a-combined-ca-bundle\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.763553 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.765214 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" event={"ID":"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b","Type":"ContainerDied","Data":"d195b2fb206399fb071c5cc8aeb1b830f25e80fbee8a3cc0133a5de92ece4c0c"} Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.765264 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d195b2fb206399fb071c5cc8aeb1b830f25e80fbee8a3cc0133a5de92ece4c0c" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.765645 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.766858 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b91c83f-0036-42a8-97b4-86846f14466a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.778159 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6w4b\" (UniqueName: \"kubernetes.io/projected/0b91c83f-0036-42a8-97b4-86846f14466a-kube-api-access-g6w4b\") pod \"ovn-controller-metrics-4kl52\" (UID: \"0b91c83f-0036-42a8-97b4-86846f14466a\") " pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.792527 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-4kl52" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.816832 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.951260 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.951890 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-dns-svc\") pod \"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b\" (UID: \"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b\") " Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.952059 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-config\") pod \"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b\" (UID: \"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b\") " Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.952241 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jq79\" (UniqueName: \"kubernetes.io/projected/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-kube-api-access-6jq79\") pod \"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b\" (UID: \"c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b\") " Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.954065 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b" (UID: "c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.957849 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-config" (OuterVolumeSpecName: "config") pod "c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b" (UID: "c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:05 crc kubenswrapper[4806]: I1204 04:13:05.962381 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-kube-api-access-6jq79" (OuterVolumeSpecName: "kube-api-access-6jq79") pod "c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b" (UID: "c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b"). InnerVolumeSpecName "kube-api-access-6jq79". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.009074 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z2hxt"] Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.054221 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jq79\" (UniqueName: \"kubernetes.io/projected/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-kube-api-access-6jq79\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.054268 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.054283 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.071381 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-xg8j9"] Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.073181 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.151831 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.164282 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-xg8j9\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.164372 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-xg8j9\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.164406 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-xg8j9\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.164467 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96skb\" (UniqueName: \"kubernetes.io/projected/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-kube-api-access-96skb\") pod \"dnsmasq-dns-86db49b7ff-xg8j9\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.164498 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-config\") pod \"dnsmasq-dns-86db49b7ff-xg8j9\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.182682 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-xg8j9"] Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.267616 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-xg8j9\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.267677 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-xg8j9\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.267758 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96skb\" (UniqueName: \"kubernetes.io/projected/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-kube-api-access-96skb\") pod \"dnsmasq-dns-86db49b7ff-xg8j9\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.267784 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-config\") pod \"dnsmasq-dns-86db49b7ff-xg8j9\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.267838 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-xg8j9\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.269900 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-config\") pod \"dnsmasq-dns-86db49b7ff-xg8j9\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.270756 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-xg8j9\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.275264 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-xg8j9\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.293915 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-xg8j9\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.323883 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96skb\" (UniqueName: \"kubernetes.io/projected/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-kube-api-access-96skb\") pod \"dnsmasq-dns-86db49b7ff-xg8j9\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.490209 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.492943 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.501476 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-p75ts" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.516554 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.516836 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.516966 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.539981 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.575849 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.580198 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqvvb\" (UniqueName: \"kubernetes.io/projected/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-kube-api-access-lqvvb\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.580293 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.580331 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.580355 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-config\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.580408 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.580437 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.580487 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-scripts\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.698550 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-scripts\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.699609 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqvvb\" (UniqueName: \"kubernetes.io/projected/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-kube-api-access-lqvvb\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.699760 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.699840 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.699891 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-config\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.700056 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.700138 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.701455 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-scripts\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.702113 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-config\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.710516 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.711033 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-xfnvf"] Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.716874 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.737827 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.758573 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.767409 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqvvb\" (UniqueName: \"kubernetes.io/projected/55f867c0-7de7-40bc-839a-4f9d9d09dd2e-kube-api-access-lqvvb\") pod \"ovn-northd-0\" (UID: \"55f867c0-7de7-40bc-839a-4f9d9d09dd2e\") " pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.798297 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" event={"ID":"711393c3-c7a7-4beb-9971-dea2c573379a","Type":"ContainerStarted","Data":"80cfd83aaf03a013735226c0a629855c862b2667dc1dc7ec2368e85c9ad9453b"} Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.803320 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-4kl52"] Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.819472 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fe13c4d3-85c5-44aa-8424-415f4a775165","Type":"ContainerStarted","Data":"f2ff17d8ba209e64829488fafa2318c3aad6b707b06987d14793cbbc7301c304"} Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.819878 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 04 04:13:06 crc kubenswrapper[4806]: W1204 04:13:06.835761 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b91c83f_0036_42a8_97b4_86846f14466a.slice/crio-501282eb685e84bf13725b6b19af302a2a5e5aafe945fe5225a744829e63e91c WatchSource:0}: Error finding container 501282eb685e84bf13725b6b19af302a2a5e5aafe945fe5225a744829e63e91c: Status 404 returned error can't find the container with id 501282eb685e84bf13725b6b19af302a2a5e5aafe945fe5225a744829e63e91c Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.835844 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" event={"ID":"3b288f07-5df3-4373-bb91-631b4cf409a4","Type":"ContainerStarted","Data":"57f99d5ad90182c4056ec1e2f176177baa21a9ef83d04e2096e26c94838323b4"} Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.836103 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" podUID="3b288f07-5df3-4373-bb91-631b4cf409a4" containerName="dnsmasq-dns" containerID="cri-o://57f99d5ad90182c4056ec1e2f176177baa21a9ef83d04e2096e26c94838323b4" gracePeriod=10 Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.836447 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.836573 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-ht488" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.851400 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.856545 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.039086716 podStartE2EDuration="42.85650605s" podCreationTimestamp="2025-12-04 04:12:24 +0000 UTC" firstStartedPulling="2025-12-04 04:12:25.165655261 +0000 UTC m=+1060.024168189" lastFinishedPulling="2025-12-04 04:13:05.983074195 +0000 UTC m=+1100.841587523" observedRunningTime="2025-12-04 04:13:06.847294688 +0000 UTC m=+1101.705807616" watchObservedRunningTime="2025-12-04 04:13:06.85650605 +0000 UTC m=+1101.715018978" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.896208 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" podStartSLOduration=2.997311062 podStartE2EDuration="48.896185034s" podCreationTimestamp="2025-12-04 04:12:18 +0000 UTC" firstStartedPulling="2025-12-04 04:12:19.160028013 +0000 UTC m=+1054.018540941" lastFinishedPulling="2025-12-04 04:13:05.058901995 +0000 UTC m=+1099.917414913" observedRunningTime="2025-12-04 04:13:06.883241375 +0000 UTC m=+1101.741754303" watchObservedRunningTime="2025-12-04 04:13:06.896185034 +0000 UTC m=+1101.754697962" Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.974239 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-ht488"] Dec 04 04:13:06 crc kubenswrapper[4806]: I1204 04:13:06.981185 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-ht488"] Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.287054 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-xg8j9"] Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.441286 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b" path="/var/lib/kubelet/pods/c1e0b1f6-f4a8-4791-8b68-bf16dfa00c8b/volumes" Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.481705 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.536967 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.644331 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcg2z\" (UniqueName: \"kubernetes.io/projected/3b288f07-5df3-4373-bb91-631b4cf409a4-kube-api-access-hcg2z\") pod \"3b288f07-5df3-4373-bb91-631b4cf409a4\" (UID: \"3b288f07-5df3-4373-bb91-631b4cf409a4\") " Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.644911 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b288f07-5df3-4373-bb91-631b4cf409a4-dns-svc\") pod \"3b288f07-5df3-4373-bb91-631b4cf409a4\" (UID: \"3b288f07-5df3-4373-bb91-631b4cf409a4\") " Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.645151 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b288f07-5df3-4373-bb91-631b4cf409a4-config\") pod \"3b288f07-5df3-4373-bb91-631b4cf409a4\" (UID: \"3b288f07-5df3-4373-bb91-631b4cf409a4\") " Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.677820 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.696792 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b288f07-5df3-4373-bb91-631b4cf409a4-kube-api-access-hcg2z" (OuterVolumeSpecName: "kube-api-access-hcg2z") pod "3b288f07-5df3-4373-bb91-631b4cf409a4" (UID: "3b288f07-5df3-4373-bb91-631b4cf409a4"). InnerVolumeSpecName "kube-api-access-hcg2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.747625 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcg2z\" (UniqueName: \"kubernetes.io/projected/3b288f07-5df3-4373-bb91-631b4cf409a4-kube-api-access-hcg2z\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.758476 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b288f07-5df3-4373-bb91-631b4cf409a4-config" (OuterVolumeSpecName: "config") pod "3b288f07-5df3-4373-bb91-631b4cf409a4" (UID: "3b288f07-5df3-4373-bb91-631b4cf409a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.798461 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b288f07-5df3-4373-bb91-631b4cf409a4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3b288f07-5df3-4373-bb91-631b4cf409a4" (UID: "3b288f07-5df3-4373-bb91-631b4cf409a4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.848861 4806 generic.go:334] "Generic (PLEG): container finished" podID="0d1a2b83-04d7-4db8-acd1-456cc0c33206" containerID="37f7b3f064bdd4c7f6ea365eccc649da0af3e90a691266d3ba657007fed82e06" exitCode=0 Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.848954 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0d1a2b83-04d7-4db8-acd1-456cc0c33206","Type":"ContainerDied","Data":"37f7b3f064bdd4c7f6ea365eccc649da0af3e90a691266d3ba657007fed82e06"} Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.850015 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b288f07-5df3-4373-bb91-631b4cf409a4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.850036 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b288f07-5df3-4373-bb91-631b4cf409a4-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.864659 4806 generic.go:334] "Generic (PLEG): container finished" podID="3b288f07-5df3-4373-bb91-631b4cf409a4" containerID="57f99d5ad90182c4056ec1e2f176177baa21a9ef83d04e2096e26c94838323b4" exitCode=0 Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.864791 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" event={"ID":"3b288f07-5df3-4373-bb91-631b4cf409a4","Type":"ContainerDied","Data":"57f99d5ad90182c4056ec1e2f176177baa21a9ef83d04e2096e26c94838323b4"} Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.864845 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" event={"ID":"3b288f07-5df3-4373-bb91-631b4cf409a4","Type":"ContainerDied","Data":"9940b55480ee4ec450a86e3e810c42a5024c27b828eb3940b040437a9579a988"} Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.864872 4806 scope.go:117] "RemoveContainer" containerID="57f99d5ad90182c4056ec1e2f176177baa21a9ef83d04e2096e26c94838323b4" Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.865116 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-z2hxt" Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.878437 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" event={"ID":"3926dc72-a9de-440b-b4ef-5a8b97bbfc31","Type":"ContainerStarted","Data":"25ae8db3734b4caf8b67539d0416bd783ecfec46c7f8ef3c37314175b283100d"} Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.900042 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" event={"ID":"711393c3-c7a7-4beb-9971-dea2c573379a","Type":"ContainerDied","Data":"85614e405a2b2cf6604b9a1e3b59f7481df752ebe88638255c18f08f6a3a299b"} Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.897954 4806 generic.go:334] "Generic (PLEG): container finished" podID="711393c3-c7a7-4beb-9971-dea2c573379a" containerID="85614e405a2b2cf6604b9a1e3b59f7481df752ebe88638255c18f08f6a3a299b" exitCode=0 Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.917143 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-4kl52" event={"ID":"0b91c83f-0036-42a8-97b4-86846f14466a","Type":"ContainerStarted","Data":"6ee132ab04e626cff7e62a19b020d098071f08d14587dfc730636bf6d7d19224"} Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.917207 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-4kl52" event={"ID":"0b91c83f-0036-42a8-97b4-86846f14466a","Type":"ContainerStarted","Data":"501282eb685e84bf13725b6b19af302a2a5e5aafe945fe5225a744829e63e91c"} Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.927129 4806 scope.go:117] "RemoveContainer" containerID="a5d4327af30e54de906adbc592216521a6ac03d0058bbab62b67ae65d3c87e8f" Dec 04 04:13:07 crc kubenswrapper[4806]: I1204 04:13:07.991216 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"55f867c0-7de7-40bc-839a-4f9d9d09dd2e","Type":"ContainerStarted","Data":"aa8eab2c0b63ee6f93c1d71f63430427da3528082e1880732c06b95541389e5e"} Dec 04 04:13:08 crc kubenswrapper[4806]: I1204 04:13:08.032471 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-4kl52" podStartSLOduration=3.032442979 podStartE2EDuration="3.032442979s" podCreationTimestamp="2025-12-04 04:13:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:13:07.992198816 +0000 UTC m=+1102.850711764" watchObservedRunningTime="2025-12-04 04:13:08.032442979 +0000 UTC m=+1102.890955907" Dec 04 04:13:08 crc kubenswrapper[4806]: I1204 04:13:08.111527 4806 scope.go:117] "RemoveContainer" containerID="57f99d5ad90182c4056ec1e2f176177baa21a9ef83d04e2096e26c94838323b4" Dec 04 04:13:08 crc kubenswrapper[4806]: I1204 04:13:08.119149 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z2hxt"] Dec 04 04:13:08 crc kubenswrapper[4806]: E1204 04:13:08.131984 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57f99d5ad90182c4056ec1e2f176177baa21a9ef83d04e2096e26c94838323b4\": container with ID starting with 57f99d5ad90182c4056ec1e2f176177baa21a9ef83d04e2096e26c94838323b4 not found: ID does not exist" containerID="57f99d5ad90182c4056ec1e2f176177baa21a9ef83d04e2096e26c94838323b4" Dec 04 04:13:08 crc kubenswrapper[4806]: I1204 04:13:08.132041 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57f99d5ad90182c4056ec1e2f176177baa21a9ef83d04e2096e26c94838323b4"} err="failed to get container status \"57f99d5ad90182c4056ec1e2f176177baa21a9ef83d04e2096e26c94838323b4\": rpc error: code = NotFound desc = could not find container \"57f99d5ad90182c4056ec1e2f176177baa21a9ef83d04e2096e26c94838323b4\": container with ID starting with 57f99d5ad90182c4056ec1e2f176177baa21a9ef83d04e2096e26c94838323b4 not found: ID does not exist" Dec 04 04:13:08 crc kubenswrapper[4806]: I1204 04:13:08.132081 4806 scope.go:117] "RemoveContainer" containerID="a5d4327af30e54de906adbc592216521a6ac03d0058bbab62b67ae65d3c87e8f" Dec 04 04:13:08 crc kubenswrapper[4806]: E1204 04:13:08.137013 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5d4327af30e54de906adbc592216521a6ac03d0058bbab62b67ae65d3c87e8f\": container with ID starting with a5d4327af30e54de906adbc592216521a6ac03d0058bbab62b67ae65d3c87e8f not found: ID does not exist" containerID="a5d4327af30e54de906adbc592216521a6ac03d0058bbab62b67ae65d3c87e8f" Dec 04 04:13:08 crc kubenswrapper[4806]: I1204 04:13:08.137112 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5d4327af30e54de906adbc592216521a6ac03d0058bbab62b67ae65d3c87e8f"} err="failed to get container status \"a5d4327af30e54de906adbc592216521a6ac03d0058bbab62b67ae65d3c87e8f\": rpc error: code = NotFound desc = could not find container \"a5d4327af30e54de906adbc592216521a6ac03d0058bbab62b67ae65d3c87e8f\": container with ID starting with a5d4327af30e54de906adbc592216521a6ac03d0058bbab62b67ae65d3c87e8f not found: ID does not exist" Dec 04 04:13:08 crc kubenswrapper[4806]: I1204 04:13:08.161812 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-z2hxt"] Dec 04 04:13:09 crc kubenswrapper[4806]: I1204 04:13:09.008550 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" event={"ID":"711393c3-c7a7-4beb-9971-dea2c573379a","Type":"ContainerStarted","Data":"afdd4a963601f6184a9e0894f57247f9fa7e007dab92d01b676a3f824494ed8a"} Dec 04 04:13:09 crc kubenswrapper[4806]: I1204 04:13:09.009093 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:09 crc kubenswrapper[4806]: I1204 04:13:09.014878 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"0d1a2b83-04d7-4db8-acd1-456cc0c33206","Type":"ContainerStarted","Data":"a9d7d414382ae1697e6e22590e0f15ac1deb8a8daa5d691b9bae8270e15cebb8"} Dec 04 04:13:09 crc kubenswrapper[4806]: I1204 04:13:09.024073 4806 generic.go:334] "Generic (PLEG): container finished" podID="3926dc72-a9de-440b-b4ef-5a8b97bbfc31" containerID="f67679577865387453b43b29d4779cf77a0490241eb49efee4ab673bc51a4de4" exitCode=0 Dec 04 04:13:09 crc kubenswrapper[4806]: I1204 04:13:09.024598 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" event={"ID":"3926dc72-a9de-440b-b4ef-5a8b97bbfc31","Type":"ContainerDied","Data":"f67679577865387453b43b29d4779cf77a0490241eb49efee4ab673bc51a4de4"} Dec 04 04:13:09 crc kubenswrapper[4806]: I1204 04:13:09.066370 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" podStartSLOduration=4.066328096 podStartE2EDuration="4.066328096s" podCreationTimestamp="2025-12-04 04:13:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:13:09.039133657 +0000 UTC m=+1103.897646605" watchObservedRunningTime="2025-12-04 04:13:09.066328096 +0000 UTC m=+1103.924841024" Dec 04 04:13:09 crc kubenswrapper[4806]: I1204 04:13:09.185164 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=-9223371987.669645 podStartE2EDuration="49.185129773s" podCreationTimestamp="2025-12-04 04:12:20 +0000 UTC" firstStartedPulling="2025-12-04 04:12:23.658886582 +0000 UTC m=+1058.517399510" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:13:09.07529649 +0000 UTC m=+1103.933809418" watchObservedRunningTime="2025-12-04 04:13:09.185129773 +0000 UTC m=+1104.043642701" Dec 04 04:13:09 crc kubenswrapper[4806]: I1204 04:13:09.439377 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b288f07-5df3-4373-bb91-631b4cf409a4" path="/var/lib/kubelet/pods/3b288f07-5df3-4373-bb91-631b4cf409a4/volumes" Dec 04 04:13:10 crc kubenswrapper[4806]: I1204 04:13:10.036523 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"55f867c0-7de7-40bc-839a-4f9d9d09dd2e","Type":"ContainerStarted","Data":"e8e426beb0266ecd868d569e3dbf630370a6e9ae99df3d87ba156a49c3e66423"} Dec 04 04:13:10 crc kubenswrapper[4806]: I1204 04:13:10.039473 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" event={"ID":"3926dc72-a9de-440b-b4ef-5a8b97bbfc31","Type":"ContainerStarted","Data":"1a165d63abfce6af4d59716f4859686eece678b6bee2aee6948f8d7e194cd40e"} Dec 04 04:13:10 crc kubenswrapper[4806]: I1204 04:13:10.068717 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" podStartSLOduration=4.068690858 podStartE2EDuration="4.068690858s" podCreationTimestamp="2025-12-04 04:13:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:13:10.063878116 +0000 UTC m=+1104.922391044" watchObservedRunningTime="2025-12-04 04:13:10.068690858 +0000 UTC m=+1104.927203786" Dec 04 04:13:11 crc kubenswrapper[4806]: I1204 04:13:11.060614 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"55f867c0-7de7-40bc-839a-4f9d9d09dd2e","Type":"ContainerStarted","Data":"a4175d60df682717c29ce43092dd02ce0a0c1eb25e2d221c25fce79e5ce01db5"} Dec 04 04:13:11 crc kubenswrapper[4806]: I1204 04:13:11.061242 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:11 crc kubenswrapper[4806]: I1204 04:13:11.114279 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.160217129 podStartE2EDuration="5.114230998s" podCreationTimestamp="2025-12-04 04:13:06 +0000 UTC" firstStartedPulling="2025-12-04 04:13:07.718773352 +0000 UTC m=+1102.577286280" lastFinishedPulling="2025-12-04 04:13:09.672787221 +0000 UTC m=+1104.531300149" observedRunningTime="2025-12-04 04:13:11.093681471 +0000 UTC m=+1105.952194399" watchObservedRunningTime="2025-12-04 04:13:11.114230998 +0000 UTC m=+1105.972743926" Dec 04 04:13:11 crc kubenswrapper[4806]: I1204 04:13:11.188627 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 04 04:13:11 crc kubenswrapper[4806]: I1204 04:13:11.188687 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 04 04:13:11 crc kubenswrapper[4806]: I1204 04:13:11.374230 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 04 04:13:11 crc kubenswrapper[4806]: I1204 04:13:11.853286 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.172656 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.465182 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-mz5hn"] Dec 04 04:13:12 crc kubenswrapper[4806]: E1204 04:13:12.465524 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b288f07-5df3-4373-bb91-631b4cf409a4" containerName="init" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.465538 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b288f07-5df3-4373-bb91-631b4cf409a4" containerName="init" Dec 04 04:13:12 crc kubenswrapper[4806]: E1204 04:13:12.465555 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b288f07-5df3-4373-bb91-631b4cf409a4" containerName="dnsmasq-dns" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.465563 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b288f07-5df3-4373-bb91-631b4cf409a4" containerName="dnsmasq-dns" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.465722 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b288f07-5df3-4373-bb91-631b4cf409a4" containerName="dnsmasq-dns" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.466243 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mz5hn" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.482657 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-mz5hn"] Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.487803 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-bd6f-account-create-update-xk87b"] Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.489037 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bd6f-account-create-update-xk87b" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.493253 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.526488 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bd6f-account-create-update-xk87b"] Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.584308 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3da131f7-72d9-4cc8-bc7b-f361b456f08c-operator-scripts\") pod \"placement-bd6f-account-create-update-xk87b\" (UID: \"3da131f7-72d9-4cc8-bc7b-f361b456f08c\") " pod="openstack/placement-bd6f-account-create-update-xk87b" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.584366 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/361a7154-0c20-4c88-bedc-f2bb0ea4d08d-operator-scripts\") pod \"placement-db-create-mz5hn\" (UID: \"361a7154-0c20-4c88-bedc-f2bb0ea4d08d\") " pod="openstack/placement-db-create-mz5hn" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.584517 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6m5m\" (UniqueName: \"kubernetes.io/projected/361a7154-0c20-4c88-bedc-f2bb0ea4d08d-kube-api-access-n6m5m\") pod \"placement-db-create-mz5hn\" (UID: \"361a7154-0c20-4c88-bedc-f2bb0ea4d08d\") " pod="openstack/placement-db-create-mz5hn" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.584551 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb9bv\" (UniqueName: \"kubernetes.io/projected/3da131f7-72d9-4cc8-bc7b-f361b456f08c-kube-api-access-vb9bv\") pod \"placement-bd6f-account-create-update-xk87b\" (UID: \"3da131f7-72d9-4cc8-bc7b-f361b456f08c\") " pod="openstack/placement-bd6f-account-create-update-xk87b" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.635224 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.635983 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.686264 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6m5m\" (UniqueName: \"kubernetes.io/projected/361a7154-0c20-4c88-bedc-f2bb0ea4d08d-kube-api-access-n6m5m\") pod \"placement-db-create-mz5hn\" (UID: \"361a7154-0c20-4c88-bedc-f2bb0ea4d08d\") " pod="openstack/placement-db-create-mz5hn" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.686335 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb9bv\" (UniqueName: \"kubernetes.io/projected/3da131f7-72d9-4cc8-bc7b-f361b456f08c-kube-api-access-vb9bv\") pod \"placement-bd6f-account-create-update-xk87b\" (UID: \"3da131f7-72d9-4cc8-bc7b-f361b456f08c\") " pod="openstack/placement-bd6f-account-create-update-xk87b" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.686383 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3da131f7-72d9-4cc8-bc7b-f361b456f08c-operator-scripts\") pod \"placement-bd6f-account-create-update-xk87b\" (UID: \"3da131f7-72d9-4cc8-bc7b-f361b456f08c\") " pod="openstack/placement-bd6f-account-create-update-xk87b" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.686399 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/361a7154-0c20-4c88-bedc-f2bb0ea4d08d-operator-scripts\") pod \"placement-db-create-mz5hn\" (UID: \"361a7154-0c20-4c88-bedc-f2bb0ea4d08d\") " pod="openstack/placement-db-create-mz5hn" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.687467 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/361a7154-0c20-4c88-bedc-f2bb0ea4d08d-operator-scripts\") pod \"placement-db-create-mz5hn\" (UID: \"361a7154-0c20-4c88-bedc-f2bb0ea4d08d\") " pod="openstack/placement-db-create-mz5hn" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.687766 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3da131f7-72d9-4cc8-bc7b-f361b456f08c-operator-scripts\") pod \"placement-bd6f-account-create-update-xk87b\" (UID: \"3da131f7-72d9-4cc8-bc7b-f361b456f08c\") " pod="openstack/placement-bd6f-account-create-update-xk87b" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.711041 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb9bv\" (UniqueName: \"kubernetes.io/projected/3da131f7-72d9-4cc8-bc7b-f361b456f08c-kube-api-access-vb9bv\") pod \"placement-bd6f-account-create-update-xk87b\" (UID: \"3da131f7-72d9-4cc8-bc7b-f361b456f08c\") " pod="openstack/placement-bd6f-account-create-update-xk87b" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.719782 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6m5m\" (UniqueName: \"kubernetes.io/projected/361a7154-0c20-4c88-bedc-f2bb0ea4d08d-kube-api-access-n6m5m\") pod \"placement-db-create-mz5hn\" (UID: \"361a7154-0c20-4c88-bedc-f2bb0ea4d08d\") " pod="openstack/placement-db-create-mz5hn" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.796708 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mz5hn" Dec 04 04:13:12 crc kubenswrapper[4806]: I1204 04:13:12.807715 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bd6f-account-create-update-xk87b" Dec 04 04:13:13 crc kubenswrapper[4806]: I1204 04:13:13.190642 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-mz5hn"] Dec 04 04:13:13 crc kubenswrapper[4806]: I1204 04:13:13.273757 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bd6f-account-create-update-xk87b"] Dec 04 04:13:13 crc kubenswrapper[4806]: I1204 04:13:13.438140 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.099746 4806 generic.go:334] "Generic (PLEG): container finished" podID="361a7154-0c20-4c88-bedc-f2bb0ea4d08d" containerID="4dd482f163c77b04bf138f255793d02c3127369b5cb07280687282da52d0e5e0" exitCode=0 Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.100386 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mz5hn" event={"ID":"361a7154-0c20-4c88-bedc-f2bb0ea4d08d","Type":"ContainerDied","Data":"4dd482f163c77b04bf138f255793d02c3127369b5cb07280687282da52d0e5e0"} Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.100470 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mz5hn" event={"ID":"361a7154-0c20-4c88-bedc-f2bb0ea4d08d","Type":"ContainerStarted","Data":"27efd5075fdd18c0eb3ecd15344926415f684bb2c10276b4dcf53b97e0c23ae2"} Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.103215 4806 generic.go:334] "Generic (PLEG): container finished" podID="3da131f7-72d9-4cc8-bc7b-f361b456f08c" containerID="d4c0f4f92bc396d80d312491155fad120886c6711e7f693b14767625c170833c" exitCode=0 Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.104606 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bd6f-account-create-update-xk87b" event={"ID":"3da131f7-72d9-4cc8-bc7b-f361b456f08c","Type":"ContainerDied","Data":"d4c0f4f92bc396d80d312491155fad120886c6711e7f693b14767625c170833c"} Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.104646 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bd6f-account-create-update-xk87b" event={"ID":"3da131f7-72d9-4cc8-bc7b-f361b456f08c","Type":"ContainerStarted","Data":"98c4d8e21b75e111c40e3b9d1688f213e1e38630589d4299383d55949f3f66c6"} Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.423133 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.441731 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-xfnvf"] Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.442174 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" podUID="711393c3-c7a7-4beb-9971-dea2c573379a" containerName="dnsmasq-dns" containerID="cri-o://afdd4a963601f6184a9e0894f57247f9fa7e007dab92d01b676a3f824494ed8a" gracePeriod=10 Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.448128 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.453431 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.500699 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-rl9g4"] Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.502951 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.588999 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-rl9g4"] Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.635511 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2nts\" (UniqueName: \"kubernetes.io/projected/00e618da-f470-4f5e-8955-f5dfd900eceb-kube-api-access-c2nts\") pod \"dnsmasq-dns-698758b865-rl9g4\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.636045 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-rl9g4\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.636125 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-dns-svc\") pod \"dnsmasq-dns-698758b865-rl9g4\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.636173 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-config\") pod \"dnsmasq-dns-698758b865-rl9g4\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.636206 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-rl9g4\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.738255 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2nts\" (UniqueName: \"kubernetes.io/projected/00e618da-f470-4f5e-8955-f5dfd900eceb-kube-api-access-c2nts\") pod \"dnsmasq-dns-698758b865-rl9g4\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.738331 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-rl9g4\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.738384 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-dns-svc\") pod \"dnsmasq-dns-698758b865-rl9g4\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.738414 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-config\") pod \"dnsmasq-dns-698758b865-rl9g4\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.738446 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-rl9g4\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.739375 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-rl9g4\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.739613 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-rl9g4\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.739694 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-dns-svc\") pod \"dnsmasq-dns-698758b865-rl9g4\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.739968 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-config\") pod \"dnsmasq-dns-698758b865-rl9g4\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.767584 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2nts\" (UniqueName: \"kubernetes.io/projected/00e618da-f470-4f5e-8955-f5dfd900eceb-kube-api-access-c2nts\") pod \"dnsmasq-dns-698758b865-rl9g4\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:14 crc kubenswrapper[4806]: I1204 04:13:14.822528 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.122409 4806 generic.go:334] "Generic (PLEG): container finished" podID="711393c3-c7a7-4beb-9971-dea2c573379a" containerID="afdd4a963601f6184a9e0894f57247f9fa7e007dab92d01b676a3f824494ed8a" exitCode=0 Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.122985 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" event={"ID":"711393c3-c7a7-4beb-9971-dea2c573379a","Type":"ContainerDied","Data":"afdd4a963601f6184a9e0894f57247f9fa7e007dab92d01b676a3f824494ed8a"} Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.372767 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-rl9g4"] Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.486366 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.499505 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 04 04:13:15 crc kubenswrapper[4806]: E1204 04:13:15.499878 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="711393c3-c7a7-4beb-9971-dea2c573379a" containerName="init" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.499900 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="711393c3-c7a7-4beb-9971-dea2c573379a" containerName="init" Dec 04 04:13:15 crc kubenswrapper[4806]: E1204 04:13:15.499982 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="711393c3-c7a7-4beb-9971-dea2c573379a" containerName="dnsmasq-dns" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.499992 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="711393c3-c7a7-4beb-9971-dea2c573379a" containerName="dnsmasq-dns" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.501146 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="711393c3-c7a7-4beb-9971-dea2c573379a" containerName="dnsmasq-dns" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.511747 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.522862 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-rjfst" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.523123 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.523136 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.524691 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.530596 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.655297 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-ovsdbserver-nb\") pod \"711393c3-c7a7-4beb-9971-dea2c573379a\" (UID: \"711393c3-c7a7-4beb-9971-dea2c573379a\") " Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.655641 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-config\") pod \"711393c3-c7a7-4beb-9971-dea2c573379a\" (UID: \"711393c3-c7a7-4beb-9971-dea2c573379a\") " Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.655685 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-dns-svc\") pod \"711393c3-c7a7-4beb-9971-dea2c573379a\" (UID: \"711393c3-c7a7-4beb-9971-dea2c573379a\") " Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.655736 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9xfp\" (UniqueName: \"kubernetes.io/projected/711393c3-c7a7-4beb-9971-dea2c573379a-kube-api-access-r9xfp\") pod \"711393c3-c7a7-4beb-9971-dea2c573379a\" (UID: \"711393c3-c7a7-4beb-9971-dea2c573379a\") " Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.656098 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/44220187-6c1e-4755-8420-d629dd59a84a-cache\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.656123 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.656178 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkqtp\" (UniqueName: \"kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-kube-api-access-kkqtp\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.656201 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.656282 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/44220187-6c1e-4755-8420-d629dd59a84a-lock\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.665059 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/711393c3-c7a7-4beb-9971-dea2c573379a-kube-api-access-r9xfp" (OuterVolumeSpecName: "kube-api-access-r9xfp") pod "711393c3-c7a7-4beb-9971-dea2c573379a" (UID: "711393c3-c7a7-4beb-9971-dea2c573379a"). InnerVolumeSpecName "kube-api-access-r9xfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.713011 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-config" (OuterVolumeSpecName: "config") pod "711393c3-c7a7-4beb-9971-dea2c573379a" (UID: "711393c3-c7a7-4beb-9971-dea2c573379a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.725193 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "711393c3-c7a7-4beb-9971-dea2c573379a" (UID: "711393c3-c7a7-4beb-9971-dea2c573379a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.732113 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bd6f-account-create-update-xk87b" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.758943 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/44220187-6c1e-4755-8420-d629dd59a84a-lock\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.759004 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/44220187-6c1e-4755-8420-d629dd59a84a-cache\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.759023 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.759111 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkqtp\" (UniqueName: \"kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-kube-api-access-kkqtp\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.759165 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.759239 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.759252 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.759266 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9xfp\" (UniqueName: \"kubernetes.io/projected/711393c3-c7a7-4beb-9971-dea2c573379a-kube-api-access-r9xfp\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.759550 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/swift-storage-0" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.759680 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/44220187-6c1e-4755-8420-d629dd59a84a-lock\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.759904 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/44220187-6c1e-4755-8420-d629dd59a84a-cache\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:15 crc kubenswrapper[4806]: E1204 04:13:15.759999 4806 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 04:13:15 crc kubenswrapper[4806]: E1204 04:13:15.760018 4806 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 04:13:15 crc kubenswrapper[4806]: E1204 04:13:15.760051 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift podName:44220187-6c1e-4755-8420-d629dd59a84a nodeName:}" failed. No retries permitted until 2025-12-04 04:13:16.260038014 +0000 UTC m=+1111.118550942 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift") pod "swift-storage-0" (UID: "44220187-6c1e-4755-8420-d629dd59a84a") : configmap "swift-ring-files" not found Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.781902 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.784074 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkqtp\" (UniqueName: \"kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-kube-api-access-kkqtp\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.816339 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "711393c3-c7a7-4beb-9971-dea2c573379a" (UID: "711393c3-c7a7-4beb-9971-dea2c573379a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.875575 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3da131f7-72d9-4cc8-bc7b-f361b456f08c-operator-scripts\") pod \"3da131f7-72d9-4cc8-bc7b-f361b456f08c\" (UID: \"3da131f7-72d9-4cc8-bc7b-f361b456f08c\") " Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.875683 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vb9bv\" (UniqueName: \"kubernetes.io/projected/3da131f7-72d9-4cc8-bc7b-f361b456f08c-kube-api-access-vb9bv\") pod \"3da131f7-72d9-4cc8-bc7b-f361b456f08c\" (UID: \"3da131f7-72d9-4cc8-bc7b-f361b456f08c\") " Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.876048 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/711393c3-c7a7-4beb-9971-dea2c573379a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.876746 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3da131f7-72d9-4cc8-bc7b-f361b456f08c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3da131f7-72d9-4cc8-bc7b-f361b456f08c" (UID: "3da131f7-72d9-4cc8-bc7b-f361b456f08c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.883143 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3da131f7-72d9-4cc8-bc7b-f361b456f08c-kube-api-access-vb9bv" (OuterVolumeSpecName: "kube-api-access-vb9bv") pod "3da131f7-72d9-4cc8-bc7b-f361b456f08c" (UID: "3da131f7-72d9-4cc8-bc7b-f361b456f08c"). InnerVolumeSpecName "kube-api-access-vb9bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.960088 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mz5hn" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.979413 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3da131f7-72d9-4cc8-bc7b-f361b456f08c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:15 crc kubenswrapper[4806]: I1204 04:13:15.979609 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vb9bv\" (UniqueName: \"kubernetes.io/projected/3da131f7-72d9-4cc8-bc7b-f361b456f08c-kube-api-access-vb9bv\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.080523 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6m5m\" (UniqueName: \"kubernetes.io/projected/361a7154-0c20-4c88-bedc-f2bb0ea4d08d-kube-api-access-n6m5m\") pod \"361a7154-0c20-4c88-bedc-f2bb0ea4d08d\" (UID: \"361a7154-0c20-4c88-bedc-f2bb0ea4d08d\") " Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.080604 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/361a7154-0c20-4c88-bedc-f2bb0ea4d08d-operator-scripts\") pod \"361a7154-0c20-4c88-bedc-f2bb0ea4d08d\" (UID: \"361a7154-0c20-4c88-bedc-f2bb0ea4d08d\") " Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.085141 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/361a7154-0c20-4c88-bedc-f2bb0ea4d08d-kube-api-access-n6m5m" (OuterVolumeSpecName: "kube-api-access-n6m5m") pod "361a7154-0c20-4c88-bedc-f2bb0ea4d08d" (UID: "361a7154-0c20-4c88-bedc-f2bb0ea4d08d"). InnerVolumeSpecName "kube-api-access-n6m5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.086189 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/361a7154-0c20-4c88-bedc-f2bb0ea4d08d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "361a7154-0c20-4c88-bedc-f2bb0ea4d08d" (UID: "361a7154-0c20-4c88-bedc-f2bb0ea4d08d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.189637 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6m5m\" (UniqueName: \"kubernetes.io/projected/361a7154-0c20-4c88-bedc-f2bb0ea4d08d-kube-api-access-n6m5m\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.190028 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/361a7154-0c20-4c88-bedc-f2bb0ea4d08d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.193137 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" event={"ID":"711393c3-c7a7-4beb-9971-dea2c573379a","Type":"ContainerDied","Data":"80cfd83aaf03a013735226c0a629855c862b2667dc1dc7ec2368e85c9ad9453b"} Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.193201 4806 scope.go:117] "RemoveContainer" containerID="afdd4a963601f6184a9e0894f57247f9fa7e007dab92d01b676a3f824494ed8a" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.195010 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-xfnvf" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.198227 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-t7k9p"] Dec 04 04:13:16 crc kubenswrapper[4806]: E1204 04:13:16.198785 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="361a7154-0c20-4c88-bedc-f2bb0ea4d08d" containerName="mariadb-database-create" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.198797 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="361a7154-0c20-4c88-bedc-f2bb0ea4d08d" containerName="mariadb-database-create" Dec 04 04:13:16 crc kubenswrapper[4806]: E1204 04:13:16.198967 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3da131f7-72d9-4cc8-bc7b-f361b456f08c" containerName="mariadb-account-create-update" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.198977 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3da131f7-72d9-4cc8-bc7b-f361b456f08c" containerName="mariadb-account-create-update" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.199274 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="361a7154-0c20-4c88-bedc-f2bb0ea4d08d" containerName="mariadb-database-create" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.199301 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3da131f7-72d9-4cc8-bc7b-f361b456f08c" containerName="mariadb-account-create-update" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.200086 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.212910 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.214076 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.223634 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.230686 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-mz5hn" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.231692 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-mz5hn" event={"ID":"361a7154-0c20-4c88-bedc-f2bb0ea4d08d","Type":"ContainerDied","Data":"27efd5075fdd18c0eb3ecd15344926415f684bb2c10276b4dcf53b97e0c23ae2"} Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.231740 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27efd5075fdd18c0eb3ecd15344926415f684bb2c10276b4dcf53b97e0c23ae2" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.254760 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-t7k9p"] Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.288803 4806 scope.go:117] "RemoveContainer" containerID="85614e405a2b2cf6604b9a1e3b59f7481df752ebe88638255c18f08f6a3a299b" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.291025 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85f9dc59-e013-464d-b305-77c85cb82f4a-scripts\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.291075 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-dispersionconf\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.291096 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/85f9dc59-e013-464d-b305-77c85cb82f4a-ring-data-devices\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.291126 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-swiftconf\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.291151 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/85f9dc59-e013-464d-b305-77c85cb82f4a-etc-swift\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.291205 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.291238 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gdpm\" (UniqueName: \"kubernetes.io/projected/85f9dc59-e013-464d-b305-77c85cb82f4a-kube-api-access-5gdpm\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.291284 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-combined-ca-bundle\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: E1204 04:13:16.291473 4806 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 04:13:16 crc kubenswrapper[4806]: E1204 04:13:16.291489 4806 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 04:13:16 crc kubenswrapper[4806]: E1204 04:13:16.291527 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift podName:44220187-6c1e-4755-8420-d629dd59a84a nodeName:}" failed. No retries permitted until 2025-12-04 04:13:17.291511576 +0000 UTC m=+1112.150024504 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift") pod "swift-storage-0" (UID: "44220187-6c1e-4755-8420-d629dd59a84a") : configmap "swift-ring-files" not found Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.292677 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bd6f-account-create-update-xk87b" event={"ID":"3da131f7-72d9-4cc8-bc7b-f361b456f08c","Type":"ContainerDied","Data":"98c4d8e21b75e111c40e3b9d1688f213e1e38630589d4299383d55949f3f66c6"} Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.292709 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98c4d8e21b75e111c40e3b9d1688f213e1e38630589d4299383d55949f3f66c6" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.292827 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bd6f-account-create-update-xk87b" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.304846 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-rl9g4" event={"ID":"00e618da-f470-4f5e-8955-f5dfd900eceb","Type":"ContainerStarted","Data":"a16b6f3ed76582a93b36d1f00cac416d552c7de94a4ff28fdf85c0b1a9adbc5c"} Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.304887 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-rl9g4" event={"ID":"00e618da-f470-4f5e-8955-f5dfd900eceb","Type":"ContainerStarted","Data":"2e05c4391486237d7e6c376064cb883a0e8f6ab13a14b7acc30be5b959154a03"} Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.392262 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-swiftconf\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.392328 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/85f9dc59-e013-464d-b305-77c85cb82f4a-etc-swift\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.392467 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gdpm\" (UniqueName: \"kubernetes.io/projected/85f9dc59-e013-464d-b305-77c85cb82f4a-kube-api-access-5gdpm\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.392522 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-combined-ca-bundle\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.392567 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85f9dc59-e013-464d-b305-77c85cb82f4a-scripts\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.392639 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-dispersionconf\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.392672 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/85f9dc59-e013-464d-b305-77c85cb82f4a-ring-data-devices\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.394326 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/85f9dc59-e013-464d-b305-77c85cb82f4a-etc-swift\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.398318 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85f9dc59-e013-464d-b305-77c85cb82f4a-scripts\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.398479 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/85f9dc59-e013-464d-b305-77c85cb82f4a-ring-data-devices\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.399736 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-combined-ca-bundle\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.401525 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-dispersionconf\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.408122 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-swiftconf\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.418756 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gdpm\" (UniqueName: \"kubernetes.io/projected/85f9dc59-e013-464d-b305-77c85cb82f4a-kube-api-access-5gdpm\") pod \"swift-ring-rebalance-t7k9p\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.419202 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-xfnvf"] Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.425768 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-xfnvf"] Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.578134 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:16 crc kubenswrapper[4806]: I1204 04:13:16.710350 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:17 crc kubenswrapper[4806]: I1204 04:13:17.248393 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-t7k9p"] Dec 04 04:13:17 crc kubenswrapper[4806]: I1204 04:13:17.309608 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:17 crc kubenswrapper[4806]: E1204 04:13:17.309823 4806 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 04:13:17 crc kubenswrapper[4806]: E1204 04:13:17.309842 4806 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 04:13:17 crc kubenswrapper[4806]: E1204 04:13:17.309892 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift podName:44220187-6c1e-4755-8420-d629dd59a84a nodeName:}" failed. No retries permitted until 2025-12-04 04:13:19.309873595 +0000 UTC m=+1114.168386523 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift") pod "swift-storage-0" (UID: "44220187-6c1e-4755-8420-d629dd59a84a") : configmap "swift-ring-files" not found Dec 04 04:13:17 crc kubenswrapper[4806]: I1204 04:13:17.315769 4806 generic.go:334] "Generic (PLEG): container finished" podID="00e618da-f470-4f5e-8955-f5dfd900eceb" containerID="a16b6f3ed76582a93b36d1f00cac416d552c7de94a4ff28fdf85c0b1a9adbc5c" exitCode=0 Dec 04 04:13:17 crc kubenswrapper[4806]: I1204 04:13:17.316037 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-rl9g4" event={"ID":"00e618da-f470-4f5e-8955-f5dfd900eceb","Type":"ContainerDied","Data":"a16b6f3ed76582a93b36d1f00cac416d552c7de94a4ff28fdf85c0b1a9adbc5c"} Dec 04 04:13:17 crc kubenswrapper[4806]: I1204 04:13:17.318067 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t7k9p" event={"ID":"85f9dc59-e013-464d-b305-77c85cb82f4a","Type":"ContainerStarted","Data":"379cbea57b2a270dca26b81f9a364b5b77567393596cde6e7bcf1c9ec2b56405"} Dec 04 04:13:17 crc kubenswrapper[4806]: I1204 04:13:17.434883 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="711393c3-c7a7-4beb-9971-dea2c573379a" path="/var/lib/kubelet/pods/711393c3-c7a7-4beb-9971-dea2c573379a/volumes" Dec 04 04:13:17 crc kubenswrapper[4806]: I1204 04:13:17.792580 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-tjvgt"] Dec 04 04:13:17 crc kubenswrapper[4806]: I1204 04:13:17.793837 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-tjvgt" Dec 04 04:13:17 crc kubenswrapper[4806]: I1204 04:13:17.802125 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-tjvgt"] Dec 04 04:13:17 crc kubenswrapper[4806]: I1204 04:13:17.923371 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcrgl\" (UniqueName: \"kubernetes.io/projected/cea7a0d7-7a20-4c9a-90fd-8bd779be02b9-kube-api-access-kcrgl\") pod \"glance-db-create-tjvgt\" (UID: \"cea7a0d7-7a20-4c9a-90fd-8bd779be02b9\") " pod="openstack/glance-db-create-tjvgt" Dec 04 04:13:17 crc kubenswrapper[4806]: I1204 04:13:17.923526 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cea7a0d7-7a20-4c9a-90fd-8bd779be02b9-operator-scripts\") pod \"glance-db-create-tjvgt\" (UID: \"cea7a0d7-7a20-4c9a-90fd-8bd779be02b9\") " pod="openstack/glance-db-create-tjvgt" Dec 04 04:13:17 crc kubenswrapper[4806]: I1204 04:13:17.952549 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-122a-account-create-update-nxhgt"] Dec 04 04:13:17 crc kubenswrapper[4806]: I1204 04:13:17.953909 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-122a-account-create-update-nxhgt" Dec 04 04:13:17 crc kubenswrapper[4806]: I1204 04:13:17.957809 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 04 04:13:17 crc kubenswrapper[4806]: I1204 04:13:17.965734 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-122a-account-create-update-nxhgt"] Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.024734 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcrgl\" (UniqueName: \"kubernetes.io/projected/cea7a0d7-7a20-4c9a-90fd-8bd779be02b9-kube-api-access-kcrgl\") pod \"glance-db-create-tjvgt\" (UID: \"cea7a0d7-7a20-4c9a-90fd-8bd779be02b9\") " pod="openstack/glance-db-create-tjvgt" Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.024821 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j66gq\" (UniqueName: \"kubernetes.io/projected/cc7a00c7-53ea-4717-89b1-1266f50590d0-kube-api-access-j66gq\") pod \"glance-122a-account-create-update-nxhgt\" (UID: \"cc7a00c7-53ea-4717-89b1-1266f50590d0\") " pod="openstack/glance-122a-account-create-update-nxhgt" Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.024899 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc7a00c7-53ea-4717-89b1-1266f50590d0-operator-scripts\") pod \"glance-122a-account-create-update-nxhgt\" (UID: \"cc7a00c7-53ea-4717-89b1-1266f50590d0\") " pod="openstack/glance-122a-account-create-update-nxhgt" Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.025045 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cea7a0d7-7a20-4c9a-90fd-8bd779be02b9-operator-scripts\") pod \"glance-db-create-tjvgt\" (UID: \"cea7a0d7-7a20-4c9a-90fd-8bd779be02b9\") " pod="openstack/glance-db-create-tjvgt" Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.025723 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cea7a0d7-7a20-4c9a-90fd-8bd779be02b9-operator-scripts\") pod \"glance-db-create-tjvgt\" (UID: \"cea7a0d7-7a20-4c9a-90fd-8bd779be02b9\") " pod="openstack/glance-db-create-tjvgt" Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.046617 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcrgl\" (UniqueName: \"kubernetes.io/projected/cea7a0d7-7a20-4c9a-90fd-8bd779be02b9-kube-api-access-kcrgl\") pod \"glance-db-create-tjvgt\" (UID: \"cea7a0d7-7a20-4c9a-90fd-8bd779be02b9\") " pod="openstack/glance-db-create-tjvgt" Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.114536 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-tjvgt" Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.127557 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j66gq\" (UniqueName: \"kubernetes.io/projected/cc7a00c7-53ea-4717-89b1-1266f50590d0-kube-api-access-j66gq\") pod \"glance-122a-account-create-update-nxhgt\" (UID: \"cc7a00c7-53ea-4717-89b1-1266f50590d0\") " pod="openstack/glance-122a-account-create-update-nxhgt" Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.127666 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc7a00c7-53ea-4717-89b1-1266f50590d0-operator-scripts\") pod \"glance-122a-account-create-update-nxhgt\" (UID: \"cc7a00c7-53ea-4717-89b1-1266f50590d0\") " pod="openstack/glance-122a-account-create-update-nxhgt" Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.135861 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc7a00c7-53ea-4717-89b1-1266f50590d0-operator-scripts\") pod \"glance-122a-account-create-update-nxhgt\" (UID: \"cc7a00c7-53ea-4717-89b1-1266f50590d0\") " pod="openstack/glance-122a-account-create-update-nxhgt" Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.153247 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j66gq\" (UniqueName: \"kubernetes.io/projected/cc7a00c7-53ea-4717-89b1-1266f50590d0-kube-api-access-j66gq\") pod \"glance-122a-account-create-update-nxhgt\" (UID: \"cc7a00c7-53ea-4717-89b1-1266f50590d0\") " pod="openstack/glance-122a-account-create-update-nxhgt" Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.279290 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-122a-account-create-update-nxhgt" Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.363688 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-rl9g4" event={"ID":"00e618da-f470-4f5e-8955-f5dfd900eceb","Type":"ContainerStarted","Data":"621825b6b9c53dbec054352c702ef78629d930aed0dabd3cfba3e004932e9d8d"} Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.363979 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.399190 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-rl9g4" podStartSLOduration=4.399164127 podStartE2EDuration="4.399164127s" podCreationTimestamp="2025-12-04 04:13:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:13:18.391120134 +0000 UTC m=+1113.249633072" watchObservedRunningTime="2025-12-04 04:13:18.399164127 +0000 UTC m=+1113.257677055" Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.565589 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-tjvgt"] Dec 04 04:13:18 crc kubenswrapper[4806]: W1204 04:13:18.576497 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcea7a0d7_7a20_4c9a_90fd_8bd779be02b9.slice/crio-0f906b79bf31f5a7fee973b3a1800769e32c14778f3daacab476b0e663a4bc37 WatchSource:0}: Error finding container 0f906b79bf31f5a7fee973b3a1800769e32c14778f3daacab476b0e663a4bc37: Status 404 returned error can't find the container with id 0f906b79bf31f5a7fee973b3a1800769e32c14778f3daacab476b0e663a4bc37 Dec 04 04:13:18 crc kubenswrapper[4806]: I1204 04:13:18.812149 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-122a-account-create-update-nxhgt"] Dec 04 04:13:18 crc kubenswrapper[4806]: W1204 04:13:18.821533 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc7a00c7_53ea_4717_89b1_1266f50590d0.slice/crio-1ff9068a7585a2ea3c2f8ed24c6d7bed5f4ddcfdac636cd866d667b8287cf137 WatchSource:0}: Error finding container 1ff9068a7585a2ea3c2f8ed24c6d7bed5f4ddcfdac636cd866d667b8287cf137: Status 404 returned error can't find the container with id 1ff9068a7585a2ea3c2f8ed24c6d7bed5f4ddcfdac636cd866d667b8287cf137 Dec 04 04:13:19 crc kubenswrapper[4806]: I1204 04:13:19.373558 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:19 crc kubenswrapper[4806]: E1204 04:13:19.373754 4806 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 04:13:19 crc kubenswrapper[4806]: E1204 04:13:19.374051 4806 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 04:13:19 crc kubenswrapper[4806]: E1204 04:13:19.374128 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift podName:44220187-6c1e-4755-8420-d629dd59a84a nodeName:}" failed. No retries permitted until 2025-12-04 04:13:23.374088609 +0000 UTC m=+1118.232601537 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift") pod "swift-storage-0" (UID: "44220187-6c1e-4755-8420-d629dd59a84a") : configmap "swift-ring-files" not found Dec 04 04:13:19 crc kubenswrapper[4806]: I1204 04:13:19.383777 4806 generic.go:334] "Generic (PLEG): container finished" podID="cea7a0d7-7a20-4c9a-90fd-8bd779be02b9" containerID="0d00e7796d93a7ceb9886b3707fee3a8baa16c9e4d7883cf1312744e209b701e" exitCode=0 Dec 04 04:13:19 crc kubenswrapper[4806]: I1204 04:13:19.383815 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-tjvgt" event={"ID":"cea7a0d7-7a20-4c9a-90fd-8bd779be02b9","Type":"ContainerDied","Data":"0d00e7796d93a7ceb9886b3707fee3a8baa16c9e4d7883cf1312744e209b701e"} Dec 04 04:13:19 crc kubenswrapper[4806]: I1204 04:13:19.383849 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-tjvgt" event={"ID":"cea7a0d7-7a20-4c9a-90fd-8bd779be02b9","Type":"ContainerStarted","Data":"0f906b79bf31f5a7fee973b3a1800769e32c14778f3daacab476b0e663a4bc37"} Dec 04 04:13:19 crc kubenswrapper[4806]: I1204 04:13:19.389896 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-122a-account-create-update-nxhgt" event={"ID":"cc7a00c7-53ea-4717-89b1-1266f50590d0","Type":"ContainerDied","Data":"4d8b7a947762eae6934fd38b651c52659cb7dec30803846fa7e3a5b167146b93"} Dec 04 04:13:19 crc kubenswrapper[4806]: I1204 04:13:19.389977 4806 generic.go:334] "Generic (PLEG): container finished" podID="cc7a00c7-53ea-4717-89b1-1266f50590d0" containerID="4d8b7a947762eae6934fd38b651c52659cb7dec30803846fa7e3a5b167146b93" exitCode=0 Dec 04 04:13:19 crc kubenswrapper[4806]: I1204 04:13:19.390280 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-122a-account-create-update-nxhgt" event={"ID":"cc7a00c7-53ea-4717-89b1-1266f50590d0","Type":"ContainerStarted","Data":"1ff9068a7585a2ea3c2f8ed24c6d7bed5f4ddcfdac636cd866d667b8287cf137"} Dec 04 04:13:21 crc kubenswrapper[4806]: I1204 04:13:21.904026 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 04 04:13:21 crc kubenswrapper[4806]: I1204 04:13:21.988285 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-wjzmn"] Dec 04 04:13:21 crc kubenswrapper[4806]: I1204 04:13:21.992236 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wjzmn" Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.001164 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-wjzmn"] Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.088367 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-9fa4-account-create-update-2pkcm"] Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.089359 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9fa4-account-create-update-2pkcm" Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.098973 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9fa4-account-create-update-2pkcm"] Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.104481 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.128698 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hd8m\" (UniqueName: \"kubernetes.io/projected/c7d42f80-97fb-4174-b256-ccf414215b24-kube-api-access-4hd8m\") pod \"keystone-db-create-wjzmn\" (UID: \"c7d42f80-97fb-4174-b256-ccf414215b24\") " pod="openstack/keystone-db-create-wjzmn" Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.128753 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7d42f80-97fb-4174-b256-ccf414215b24-operator-scripts\") pod \"keystone-db-create-wjzmn\" (UID: \"c7d42f80-97fb-4174-b256-ccf414215b24\") " pod="openstack/keystone-db-create-wjzmn" Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.229850 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjvwp\" (UniqueName: \"kubernetes.io/projected/a872a38c-7d3d-4685-aed2-b224c9f18a4a-kube-api-access-mjvwp\") pod \"keystone-9fa4-account-create-update-2pkcm\" (UID: \"a872a38c-7d3d-4685-aed2-b224c9f18a4a\") " pod="openstack/keystone-9fa4-account-create-update-2pkcm" Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.229915 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hd8m\" (UniqueName: \"kubernetes.io/projected/c7d42f80-97fb-4174-b256-ccf414215b24-kube-api-access-4hd8m\") pod \"keystone-db-create-wjzmn\" (UID: \"c7d42f80-97fb-4174-b256-ccf414215b24\") " pod="openstack/keystone-db-create-wjzmn" Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.229963 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7d42f80-97fb-4174-b256-ccf414215b24-operator-scripts\") pod \"keystone-db-create-wjzmn\" (UID: \"c7d42f80-97fb-4174-b256-ccf414215b24\") " pod="openstack/keystone-db-create-wjzmn" Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.230034 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a872a38c-7d3d-4685-aed2-b224c9f18a4a-operator-scripts\") pod \"keystone-9fa4-account-create-update-2pkcm\" (UID: \"a872a38c-7d3d-4685-aed2-b224c9f18a4a\") " pod="openstack/keystone-9fa4-account-create-update-2pkcm" Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.231478 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7d42f80-97fb-4174-b256-ccf414215b24-operator-scripts\") pod \"keystone-db-create-wjzmn\" (UID: \"c7d42f80-97fb-4174-b256-ccf414215b24\") " pod="openstack/keystone-db-create-wjzmn" Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.253573 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hd8m\" (UniqueName: \"kubernetes.io/projected/c7d42f80-97fb-4174-b256-ccf414215b24-kube-api-access-4hd8m\") pod \"keystone-db-create-wjzmn\" (UID: \"c7d42f80-97fb-4174-b256-ccf414215b24\") " pod="openstack/keystone-db-create-wjzmn" Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.314438 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wjzmn" Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.331386 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a872a38c-7d3d-4685-aed2-b224c9f18a4a-operator-scripts\") pod \"keystone-9fa4-account-create-update-2pkcm\" (UID: \"a872a38c-7d3d-4685-aed2-b224c9f18a4a\") " pod="openstack/keystone-9fa4-account-create-update-2pkcm" Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.331534 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjvwp\" (UniqueName: \"kubernetes.io/projected/a872a38c-7d3d-4685-aed2-b224c9f18a4a-kube-api-access-mjvwp\") pod \"keystone-9fa4-account-create-update-2pkcm\" (UID: \"a872a38c-7d3d-4685-aed2-b224c9f18a4a\") " pod="openstack/keystone-9fa4-account-create-update-2pkcm" Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.332565 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a872a38c-7d3d-4685-aed2-b224c9f18a4a-operator-scripts\") pod \"keystone-9fa4-account-create-update-2pkcm\" (UID: \"a872a38c-7d3d-4685-aed2-b224c9f18a4a\") " pod="openstack/keystone-9fa4-account-create-update-2pkcm" Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.348775 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjvwp\" (UniqueName: \"kubernetes.io/projected/a872a38c-7d3d-4685-aed2-b224c9f18a4a-kube-api-access-mjvwp\") pod \"keystone-9fa4-account-create-update-2pkcm\" (UID: \"a872a38c-7d3d-4685-aed2-b224c9f18a4a\") " pod="openstack/keystone-9fa4-account-create-update-2pkcm" Dec 04 04:13:22 crc kubenswrapper[4806]: I1204 04:13:22.413821 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9fa4-account-create-update-2pkcm" Dec 04 04:13:23 crc kubenswrapper[4806]: I1204 04:13:23.452535 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:23 crc kubenswrapper[4806]: E1204 04:13:23.452811 4806 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 04:13:23 crc kubenswrapper[4806]: E1204 04:13:23.452830 4806 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 04:13:23 crc kubenswrapper[4806]: E1204 04:13:23.452982 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift podName:44220187-6c1e-4755-8420-d629dd59a84a nodeName:}" failed. No retries permitted until 2025-12-04 04:13:31.452965299 +0000 UTC m=+1126.311478227 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift") pod "swift-storage-0" (UID: "44220187-6c1e-4755-8420-d629dd59a84a") : configmap "swift-ring-files" not found Dec 04 04:13:23 crc kubenswrapper[4806]: I1204 04:13:23.982197 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-122a-account-create-update-nxhgt" Dec 04 04:13:23 crc kubenswrapper[4806]: I1204 04:13:23.999233 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-tjvgt" Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.061062 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc7a00c7-53ea-4717-89b1-1266f50590d0-operator-scripts\") pod \"cc7a00c7-53ea-4717-89b1-1266f50590d0\" (UID: \"cc7a00c7-53ea-4717-89b1-1266f50590d0\") " Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.061231 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j66gq\" (UniqueName: \"kubernetes.io/projected/cc7a00c7-53ea-4717-89b1-1266f50590d0-kube-api-access-j66gq\") pod \"cc7a00c7-53ea-4717-89b1-1266f50590d0\" (UID: \"cc7a00c7-53ea-4717-89b1-1266f50590d0\") " Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.061782 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc7a00c7-53ea-4717-89b1-1266f50590d0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cc7a00c7-53ea-4717-89b1-1266f50590d0" (UID: "cc7a00c7-53ea-4717-89b1-1266f50590d0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.072124 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc7a00c7-53ea-4717-89b1-1266f50590d0-kube-api-access-j66gq" (OuterVolumeSpecName: "kube-api-access-j66gq") pod "cc7a00c7-53ea-4717-89b1-1266f50590d0" (UID: "cc7a00c7-53ea-4717-89b1-1266f50590d0"). InnerVolumeSpecName "kube-api-access-j66gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.162644 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcrgl\" (UniqueName: \"kubernetes.io/projected/cea7a0d7-7a20-4c9a-90fd-8bd779be02b9-kube-api-access-kcrgl\") pod \"cea7a0d7-7a20-4c9a-90fd-8bd779be02b9\" (UID: \"cea7a0d7-7a20-4c9a-90fd-8bd779be02b9\") " Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.163241 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cea7a0d7-7a20-4c9a-90fd-8bd779be02b9-operator-scripts\") pod \"cea7a0d7-7a20-4c9a-90fd-8bd779be02b9\" (UID: \"cea7a0d7-7a20-4c9a-90fd-8bd779be02b9\") " Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.163681 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc7a00c7-53ea-4717-89b1-1266f50590d0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.163706 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j66gq\" (UniqueName: \"kubernetes.io/projected/cc7a00c7-53ea-4717-89b1-1266f50590d0-kube-api-access-j66gq\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.164108 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cea7a0d7-7a20-4c9a-90fd-8bd779be02b9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cea7a0d7-7a20-4c9a-90fd-8bd779be02b9" (UID: "cea7a0d7-7a20-4c9a-90fd-8bd779be02b9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.167651 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cea7a0d7-7a20-4c9a-90fd-8bd779be02b9-kube-api-access-kcrgl" (OuterVolumeSpecName: "kube-api-access-kcrgl") pod "cea7a0d7-7a20-4c9a-90fd-8bd779be02b9" (UID: "cea7a0d7-7a20-4c9a-90fd-8bd779be02b9"). InnerVolumeSpecName "kube-api-access-kcrgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.265191 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cea7a0d7-7a20-4c9a-90fd-8bd779be02b9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.265230 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcrgl\" (UniqueName: \"kubernetes.io/projected/cea7a0d7-7a20-4c9a-90fd-8bd779be02b9-kube-api-access-kcrgl\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.342795 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-wjzmn"] Dec 04 04:13:24 crc kubenswrapper[4806]: W1204 04:13:24.346729 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7d42f80_97fb_4174_b256_ccf414215b24.slice/crio-9582c254ac7712ff9634d9a80ead929d3642eb197341730f60b4a0debd49e225 WatchSource:0}: Error finding container 9582c254ac7712ff9634d9a80ead929d3642eb197341730f60b4a0debd49e225: Status 404 returned error can't find the container with id 9582c254ac7712ff9634d9a80ead929d3642eb197341730f60b4a0debd49e225 Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.455303 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9fa4-account-create-update-2pkcm"] Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.459480 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t7k9p" event={"ID":"85f9dc59-e013-464d-b305-77c85cb82f4a","Type":"ContainerStarted","Data":"c3b495d2aebee31942caf47530245a5afa26f0bc262a7c71fa266c7de2ae1a71"} Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.466304 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-tjvgt" Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.465842 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-tjvgt" event={"ID":"cea7a0d7-7a20-4c9a-90fd-8bd779be02b9","Type":"ContainerDied","Data":"0f906b79bf31f5a7fee973b3a1800769e32c14778f3daacab476b0e663a4bc37"} Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.469220 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f906b79bf31f5a7fee973b3a1800769e32c14778f3daacab476b0e663a4bc37" Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.485997 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-t7k9p" podStartSLOduration=1.863886454 podStartE2EDuration="8.485973719s" podCreationTimestamp="2025-12-04 04:13:16 +0000 UTC" firstStartedPulling="2025-12-04 04:13:17.24709608 +0000 UTC m=+1112.105609018" lastFinishedPulling="2025-12-04 04:13:23.869183355 +0000 UTC m=+1118.727696283" observedRunningTime="2025-12-04 04:13:24.481152238 +0000 UTC m=+1119.339665166" watchObservedRunningTime="2025-12-04 04:13:24.485973719 +0000 UTC m=+1119.344486647" Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.487650 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wjzmn" event={"ID":"c7d42f80-97fb-4174-b256-ccf414215b24","Type":"ContainerStarted","Data":"9582c254ac7712ff9634d9a80ead929d3642eb197341730f60b4a0debd49e225"} Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.491701 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-122a-account-create-update-nxhgt" Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.491556 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-122a-account-create-update-nxhgt" event={"ID":"cc7a00c7-53ea-4717-89b1-1266f50590d0","Type":"ContainerDied","Data":"1ff9068a7585a2ea3c2f8ed24c6d7bed5f4ddcfdac636cd866d667b8287cf137"} Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.493360 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ff9068a7585a2ea3c2f8ed24c6d7bed5f4ddcfdac636cd866d667b8287cf137" Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.825319 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.921548 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-xg8j9"] Dec 04 04:13:24 crc kubenswrapper[4806]: I1204 04:13:24.921837 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" podUID="3926dc72-a9de-440b-b4ef-5a8b97bbfc31" containerName="dnsmasq-dns" containerID="cri-o://1a165d63abfce6af4d59716f4859686eece678b6bee2aee6948f8d7e194cd40e" gracePeriod=10 Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.518716 4806 generic.go:334] "Generic (PLEG): container finished" podID="a872a38c-7d3d-4685-aed2-b224c9f18a4a" containerID="be588154a943039e689938cd9ed3e06c2255fef9793cb887b6feb992cd644828" exitCode=0 Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.518869 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9fa4-account-create-update-2pkcm" event={"ID":"a872a38c-7d3d-4685-aed2-b224c9f18a4a","Type":"ContainerDied","Data":"be588154a943039e689938cd9ed3e06c2255fef9793cb887b6feb992cd644828"} Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.518911 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9fa4-account-create-update-2pkcm" event={"ID":"a872a38c-7d3d-4685-aed2-b224c9f18a4a","Type":"ContainerStarted","Data":"7fddd284032ae38634de4cc81670936a26053ebaa241a25a57dcd000b35d7b2b"} Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.525870 4806 generic.go:334] "Generic (PLEG): container finished" podID="c7d42f80-97fb-4174-b256-ccf414215b24" containerID="33e9f1443117789b81d7d1a6d67bc35ac21a29cf81a9a9af687cbfd2b17e9eaf" exitCode=0 Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.526091 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wjzmn" event={"ID":"c7d42f80-97fb-4174-b256-ccf414215b24","Type":"ContainerDied","Data":"33e9f1443117789b81d7d1a6d67bc35ac21a29cf81a9a9af687cbfd2b17e9eaf"} Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.541443 4806 generic.go:334] "Generic (PLEG): container finished" podID="3926dc72-a9de-440b-b4ef-5a8b97bbfc31" containerID="1a165d63abfce6af4d59716f4859686eece678b6bee2aee6948f8d7e194cd40e" exitCode=0 Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.541579 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" event={"ID":"3926dc72-a9de-440b-b4ef-5a8b97bbfc31","Type":"ContainerDied","Data":"1a165d63abfce6af4d59716f4859686eece678b6bee2aee6948f8d7e194cd40e"} Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.601391 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.706727 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-dns-svc\") pod \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.706809 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-ovsdbserver-sb\") pod \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.706858 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96skb\" (UniqueName: \"kubernetes.io/projected/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-kube-api-access-96skb\") pod \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.706936 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-config\") pod \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.707023 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-ovsdbserver-nb\") pod \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\" (UID: \"3926dc72-a9de-440b-b4ef-5a8b97bbfc31\") " Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.730484 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-kube-api-access-96skb" (OuterVolumeSpecName: "kube-api-access-96skb") pod "3926dc72-a9de-440b-b4ef-5a8b97bbfc31" (UID: "3926dc72-a9de-440b-b4ef-5a8b97bbfc31"). InnerVolumeSpecName "kube-api-access-96skb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.768796 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3926dc72-a9de-440b-b4ef-5a8b97bbfc31" (UID: "3926dc72-a9de-440b-b4ef-5a8b97bbfc31"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.776741 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3926dc72-a9de-440b-b4ef-5a8b97bbfc31" (UID: "3926dc72-a9de-440b-b4ef-5a8b97bbfc31"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.802855 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-config" (OuterVolumeSpecName: "config") pod "3926dc72-a9de-440b-b4ef-5a8b97bbfc31" (UID: "3926dc72-a9de-440b-b4ef-5a8b97bbfc31"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.809269 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.809301 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.809319 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.809338 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96skb\" (UniqueName: \"kubernetes.io/projected/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-kube-api-access-96skb\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.825280 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3926dc72-a9de-440b-b4ef-5a8b97bbfc31" (UID: "3926dc72-a9de-440b-b4ef-5a8b97bbfc31"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:25 crc kubenswrapper[4806]: I1204 04:13:25.911281 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3926dc72-a9de-440b-b4ef-5a8b97bbfc31-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:26 crc kubenswrapper[4806]: I1204 04:13:26.554158 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" Dec 04 04:13:26 crc kubenswrapper[4806]: I1204 04:13:26.554335 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-xg8j9" event={"ID":"3926dc72-a9de-440b-b4ef-5a8b97bbfc31","Type":"ContainerDied","Data":"25ae8db3734b4caf8b67539d0416bd783ecfec46c7f8ef3c37314175b283100d"} Dec 04 04:13:26 crc kubenswrapper[4806]: I1204 04:13:26.556486 4806 scope.go:117] "RemoveContainer" containerID="1a165d63abfce6af4d59716f4859686eece678b6bee2aee6948f8d7e194cd40e" Dec 04 04:13:26 crc kubenswrapper[4806]: I1204 04:13:26.611063 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-xg8j9"] Dec 04 04:13:26 crc kubenswrapper[4806]: I1204 04:13:26.615441 4806 scope.go:117] "RemoveContainer" containerID="f67679577865387453b43b29d4779cf77a0490241eb49efee4ab673bc51a4de4" Dec 04 04:13:26 crc kubenswrapper[4806]: I1204 04:13:26.617373 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-xg8j9"] Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.000715 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9fa4-account-create-update-2pkcm" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.008622 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wjzmn" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.047414 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.047485 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.047541 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.048312 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ff51617027055d5b0c7e244b302f5ed7ba3e3cbe117dc54d0ea82947af1a87f2"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.048374 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://ff51617027055d5b0c7e244b302f5ed7ba3e3cbe117dc54d0ea82947af1a87f2" gracePeriod=600 Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.141239 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7d42f80-97fb-4174-b256-ccf414215b24-operator-scripts\") pod \"c7d42f80-97fb-4174-b256-ccf414215b24\" (UID: \"c7d42f80-97fb-4174-b256-ccf414215b24\") " Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.141345 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a872a38c-7d3d-4685-aed2-b224c9f18a4a-operator-scripts\") pod \"a872a38c-7d3d-4685-aed2-b224c9f18a4a\" (UID: \"a872a38c-7d3d-4685-aed2-b224c9f18a4a\") " Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.141451 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hd8m\" (UniqueName: \"kubernetes.io/projected/c7d42f80-97fb-4174-b256-ccf414215b24-kube-api-access-4hd8m\") pod \"c7d42f80-97fb-4174-b256-ccf414215b24\" (UID: \"c7d42f80-97fb-4174-b256-ccf414215b24\") " Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.141479 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjvwp\" (UniqueName: \"kubernetes.io/projected/a872a38c-7d3d-4685-aed2-b224c9f18a4a-kube-api-access-mjvwp\") pod \"a872a38c-7d3d-4685-aed2-b224c9f18a4a\" (UID: \"a872a38c-7d3d-4685-aed2-b224c9f18a4a\") " Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.142727 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7d42f80-97fb-4174-b256-ccf414215b24-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c7d42f80-97fb-4174-b256-ccf414215b24" (UID: "c7d42f80-97fb-4174-b256-ccf414215b24"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.142907 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a872a38c-7d3d-4685-aed2-b224c9f18a4a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a872a38c-7d3d-4685-aed2-b224c9f18a4a" (UID: "a872a38c-7d3d-4685-aed2-b224c9f18a4a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.161539 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7d42f80-97fb-4174-b256-ccf414215b24-kube-api-access-4hd8m" (OuterVolumeSpecName: "kube-api-access-4hd8m") pod "c7d42f80-97fb-4174-b256-ccf414215b24" (UID: "c7d42f80-97fb-4174-b256-ccf414215b24"). InnerVolumeSpecName "kube-api-access-4hd8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.162224 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a872a38c-7d3d-4685-aed2-b224c9f18a4a-kube-api-access-mjvwp" (OuterVolumeSpecName: "kube-api-access-mjvwp") pod "a872a38c-7d3d-4685-aed2-b224c9f18a4a" (UID: "a872a38c-7d3d-4685-aed2-b224c9f18a4a"). InnerVolumeSpecName "kube-api-access-mjvwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.246064 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7d42f80-97fb-4174-b256-ccf414215b24-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.246122 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a872a38c-7d3d-4685-aed2-b224c9f18a4a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.246134 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hd8m\" (UniqueName: \"kubernetes.io/projected/c7d42f80-97fb-4174-b256-ccf414215b24-kube-api-access-4hd8m\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.246144 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjvwp\" (UniqueName: \"kubernetes.io/projected/a872a38c-7d3d-4685-aed2-b224c9f18a4a-kube-api-access-mjvwp\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.440093 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3926dc72-a9de-440b-b4ef-5a8b97bbfc31" path="/var/lib/kubelet/pods/3926dc72-a9de-440b-b4ef-5a8b97bbfc31/volumes" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.567279 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9fa4-account-create-update-2pkcm" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.567278 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9fa4-account-create-update-2pkcm" event={"ID":"a872a38c-7d3d-4685-aed2-b224c9f18a4a","Type":"ContainerDied","Data":"7fddd284032ae38634de4cc81670936a26053ebaa241a25a57dcd000b35d7b2b"} Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.567623 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fddd284032ae38634de4cc81670936a26053ebaa241a25a57dcd000b35d7b2b" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.569188 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-wjzmn" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.569779 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-wjzmn" event={"ID":"c7d42f80-97fb-4174-b256-ccf414215b24","Type":"ContainerDied","Data":"9582c254ac7712ff9634d9a80ead929d3642eb197341730f60b4a0debd49e225"} Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.570465 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9582c254ac7712ff9634d9a80ead929d3642eb197341730f60b4a0debd49e225" Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.574971 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="ff51617027055d5b0c7e244b302f5ed7ba3e3cbe117dc54d0ea82947af1a87f2" exitCode=0 Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.575155 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"ff51617027055d5b0c7e244b302f5ed7ba3e3cbe117dc54d0ea82947af1a87f2"} Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.575194 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"95ac262c2400e25c649618e673d25b593b945bc15ad472113e55f250a2d5fdb0"} Dec 04 04:13:27 crc kubenswrapper[4806]: I1204 04:13:27.575217 4806 scope.go:117] "RemoveContainer" containerID="4ce73f616eb3f88f0847c6436643d27fc2337de01de1b065778733e8ceef4522" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.258231 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-r7mb9"] Dec 04 04:13:28 crc kubenswrapper[4806]: E1204 04:13:28.259261 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a872a38c-7d3d-4685-aed2-b224c9f18a4a" containerName="mariadb-account-create-update" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.259287 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a872a38c-7d3d-4685-aed2-b224c9f18a4a" containerName="mariadb-account-create-update" Dec 04 04:13:28 crc kubenswrapper[4806]: E1204 04:13:28.259306 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7d42f80-97fb-4174-b256-ccf414215b24" containerName="mariadb-database-create" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.259316 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7d42f80-97fb-4174-b256-ccf414215b24" containerName="mariadb-database-create" Dec 04 04:13:28 crc kubenswrapper[4806]: E1204 04:13:28.259338 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3926dc72-a9de-440b-b4ef-5a8b97bbfc31" containerName="init" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.259346 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3926dc72-a9de-440b-b4ef-5a8b97bbfc31" containerName="init" Dec 04 04:13:28 crc kubenswrapper[4806]: E1204 04:13:28.259359 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cea7a0d7-7a20-4c9a-90fd-8bd779be02b9" containerName="mariadb-database-create" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.259366 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="cea7a0d7-7a20-4c9a-90fd-8bd779be02b9" containerName="mariadb-database-create" Dec 04 04:13:28 crc kubenswrapper[4806]: E1204 04:13:28.259386 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3926dc72-a9de-440b-b4ef-5a8b97bbfc31" containerName="dnsmasq-dns" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.259395 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3926dc72-a9de-440b-b4ef-5a8b97bbfc31" containerName="dnsmasq-dns" Dec 04 04:13:28 crc kubenswrapper[4806]: E1204 04:13:28.259407 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc7a00c7-53ea-4717-89b1-1266f50590d0" containerName="mariadb-account-create-update" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.259413 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc7a00c7-53ea-4717-89b1-1266f50590d0" containerName="mariadb-account-create-update" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.259664 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="a872a38c-7d3d-4685-aed2-b224c9f18a4a" containerName="mariadb-account-create-update" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.259677 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="cea7a0d7-7a20-4c9a-90fd-8bd779be02b9" containerName="mariadb-database-create" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.259689 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3926dc72-a9de-440b-b4ef-5a8b97bbfc31" containerName="dnsmasq-dns" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.259701 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7d42f80-97fb-4174-b256-ccf414215b24" containerName="mariadb-database-create" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.259713 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc7a00c7-53ea-4717-89b1-1266f50590d0" containerName="mariadb-account-create-update" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.264761 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-r7mb9" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.268275 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.287505 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-zbhs5" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.298560 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-r7mb9"] Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.368865 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdtwf\" (UniqueName: \"kubernetes.io/projected/277e6177-7249-4994-9228-c697d5eecff1-kube-api-access-rdtwf\") pod \"glance-db-sync-r7mb9\" (UID: \"277e6177-7249-4994-9228-c697d5eecff1\") " pod="openstack/glance-db-sync-r7mb9" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.369092 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-combined-ca-bundle\") pod \"glance-db-sync-r7mb9\" (UID: \"277e6177-7249-4994-9228-c697d5eecff1\") " pod="openstack/glance-db-sync-r7mb9" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.369120 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-db-sync-config-data\") pod \"glance-db-sync-r7mb9\" (UID: \"277e6177-7249-4994-9228-c697d5eecff1\") " pod="openstack/glance-db-sync-r7mb9" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.369148 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-config-data\") pod \"glance-db-sync-r7mb9\" (UID: \"277e6177-7249-4994-9228-c697d5eecff1\") " pod="openstack/glance-db-sync-r7mb9" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.472117 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-combined-ca-bundle\") pod \"glance-db-sync-r7mb9\" (UID: \"277e6177-7249-4994-9228-c697d5eecff1\") " pod="openstack/glance-db-sync-r7mb9" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.472201 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-db-sync-config-data\") pod \"glance-db-sync-r7mb9\" (UID: \"277e6177-7249-4994-9228-c697d5eecff1\") " pod="openstack/glance-db-sync-r7mb9" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.472248 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-config-data\") pod \"glance-db-sync-r7mb9\" (UID: \"277e6177-7249-4994-9228-c697d5eecff1\") " pod="openstack/glance-db-sync-r7mb9" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.472321 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdtwf\" (UniqueName: \"kubernetes.io/projected/277e6177-7249-4994-9228-c697d5eecff1-kube-api-access-rdtwf\") pod \"glance-db-sync-r7mb9\" (UID: \"277e6177-7249-4994-9228-c697d5eecff1\") " pod="openstack/glance-db-sync-r7mb9" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.481207 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-combined-ca-bundle\") pod \"glance-db-sync-r7mb9\" (UID: \"277e6177-7249-4994-9228-c697d5eecff1\") " pod="openstack/glance-db-sync-r7mb9" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.484655 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-config-data\") pod \"glance-db-sync-r7mb9\" (UID: \"277e6177-7249-4994-9228-c697d5eecff1\") " pod="openstack/glance-db-sync-r7mb9" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.498339 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdtwf\" (UniqueName: \"kubernetes.io/projected/277e6177-7249-4994-9228-c697d5eecff1-kube-api-access-rdtwf\") pod \"glance-db-sync-r7mb9\" (UID: \"277e6177-7249-4994-9228-c697d5eecff1\") " pod="openstack/glance-db-sync-r7mb9" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.501257 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-db-sync-config-data\") pod \"glance-db-sync-r7mb9\" (UID: \"277e6177-7249-4994-9228-c697d5eecff1\") " pod="openstack/glance-db-sync-r7mb9" Dec 04 04:13:28 crc kubenswrapper[4806]: I1204 04:13:28.598645 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-r7mb9" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.208337 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-tx487" podUID="ef210587-097b-48d8-8097-dc1b95e2456e" containerName="ovn-controller" probeResult="failure" output=< Dec 04 04:13:29 crc kubenswrapper[4806]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 04 04:13:29 crc kubenswrapper[4806]: > Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.246182 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.266691 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-2sz9t" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.280578 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-r7mb9"] Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.555632 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-tx487-config-7rm8m"] Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.557102 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.561053 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.580127 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tx487-config-7rm8m"] Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.607287 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-r7mb9" event={"ID":"277e6177-7249-4994-9228-c697d5eecff1","Type":"ContainerStarted","Data":"01a2d33ae80f60f8f127060b5c6a09d4bd318cf03118146ee26760773d09dda0"} Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.607794 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-run-ovn\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.607843 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/224d1827-61ff-4882-901d-3aab54a7d598-additional-scripts\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.607876 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/224d1827-61ff-4882-901d-3aab54a7d598-scripts\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.607913 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-run\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.607949 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-log-ovn\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.607986 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59bqd\" (UniqueName: \"kubernetes.io/projected/224d1827-61ff-4882-901d-3aab54a7d598-kube-api-access-59bqd\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.610233 4806 generic.go:334] "Generic (PLEG): container finished" podID="0b7d7449-e532-4a32-bfcd-dc9f92767bd6" containerID="e59290c926c0f12d51a460b02e72a020b41534db66e0140ba00eddb2776df964" exitCode=0 Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.611071 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0b7d7449-e532-4a32-bfcd-dc9f92767bd6","Type":"ContainerDied","Data":"e59290c926c0f12d51a460b02e72a020b41534db66e0140ba00eddb2776df964"} Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.713832 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/224d1827-61ff-4882-901d-3aab54a7d598-scripts\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.714436 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-run\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.714473 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-log-ovn\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.714553 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59bqd\" (UniqueName: \"kubernetes.io/projected/224d1827-61ff-4882-901d-3aab54a7d598-kube-api-access-59bqd\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.714645 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-run-ovn\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.714708 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/224d1827-61ff-4882-901d-3aab54a7d598-additional-scripts\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.715298 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-run\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.715664 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/224d1827-61ff-4882-901d-3aab54a7d598-additional-scripts\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.715737 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-run-ovn\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.715787 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-log-ovn\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.717554 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/224d1827-61ff-4882-901d-3aab54a7d598-scripts\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.858632 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59bqd\" (UniqueName: \"kubernetes.io/projected/224d1827-61ff-4882-901d-3aab54a7d598-kube-api-access-59bqd\") pod \"ovn-controller-tx487-config-7rm8m\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:29 crc kubenswrapper[4806]: I1204 04:13:29.885192 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:30 crc kubenswrapper[4806]: I1204 04:13:30.482425 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-tx487-config-7rm8m"] Dec 04 04:13:30 crc kubenswrapper[4806]: W1204 04:13:30.497554 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod224d1827_61ff_4882_901d_3aab54a7d598.slice/crio-ebfcdc14dc6ff6ee051d19b403ebe417f6eb1035bb485bcaae058fe648c2e06c WatchSource:0}: Error finding container ebfcdc14dc6ff6ee051d19b403ebe417f6eb1035bb485bcaae058fe648c2e06c: Status 404 returned error can't find the container with id ebfcdc14dc6ff6ee051d19b403ebe417f6eb1035bb485bcaae058fe648c2e06c Dec 04 04:13:30 crc kubenswrapper[4806]: I1204 04:13:30.656567 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0b7d7449-e532-4a32-bfcd-dc9f92767bd6","Type":"ContainerStarted","Data":"fa63f5e0190e9d9dd4e1c9e13a341601eb4661dec32b87ddf56918276f9bad62"} Dec 04 04:13:30 crc kubenswrapper[4806]: I1204 04:13:30.657096 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 04 04:13:30 crc kubenswrapper[4806]: I1204 04:13:30.663438 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tx487-config-7rm8m" event={"ID":"224d1827-61ff-4882-901d-3aab54a7d598","Type":"ContainerStarted","Data":"ebfcdc14dc6ff6ee051d19b403ebe417f6eb1035bb485bcaae058fe648c2e06c"} Dec 04 04:13:30 crc kubenswrapper[4806]: I1204 04:13:30.697522 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.385811677 podStartE2EDuration="1m12.697503287s" podCreationTimestamp="2025-12-04 04:12:18 +0000 UTC" firstStartedPulling="2025-12-04 04:12:20.61488561 +0000 UTC m=+1055.473398538" lastFinishedPulling="2025-12-04 04:12:55.92657722 +0000 UTC m=+1090.785090148" observedRunningTime="2025-12-04 04:13:30.695775672 +0000 UTC m=+1125.554288600" watchObservedRunningTime="2025-12-04 04:13:30.697503287 +0000 UTC m=+1125.556016215" Dec 04 04:13:31 crc kubenswrapper[4806]: I1204 04:13:31.468074 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:31 crc kubenswrapper[4806]: E1204 04:13:31.468361 4806 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 04:13:31 crc kubenswrapper[4806]: E1204 04:13:31.468624 4806 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 04:13:31 crc kubenswrapper[4806]: E1204 04:13:31.468734 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift podName:44220187-6c1e-4755-8420-d629dd59a84a nodeName:}" failed. No retries permitted until 2025-12-04 04:13:47.46870582 +0000 UTC m=+1142.327218748 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift") pod "swift-storage-0" (UID: "44220187-6c1e-4755-8420-d629dd59a84a") : configmap "swift-ring-files" not found Dec 04 04:13:31 crc kubenswrapper[4806]: I1204 04:13:31.678333 4806 generic.go:334] "Generic (PLEG): container finished" podID="224d1827-61ff-4882-901d-3aab54a7d598" containerID="5e8e3cdd2cdc41f32edc8768a9ed26b1a0264603e7c50dbcdba6f2059cd35412" exitCode=0 Dec 04 04:13:31 crc kubenswrapper[4806]: I1204 04:13:31.678482 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tx487-config-7rm8m" event={"ID":"224d1827-61ff-4882-901d-3aab54a7d598","Type":"ContainerDied","Data":"5e8e3cdd2cdc41f32edc8768a9ed26b1a0264603e7c50dbcdba6f2059cd35412"} Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.142761 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.207689 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59bqd\" (UniqueName: \"kubernetes.io/projected/224d1827-61ff-4882-901d-3aab54a7d598-kube-api-access-59bqd\") pod \"224d1827-61ff-4882-901d-3aab54a7d598\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.208271 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/224d1827-61ff-4882-901d-3aab54a7d598-additional-scripts\") pod \"224d1827-61ff-4882-901d-3aab54a7d598\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.208315 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/224d1827-61ff-4882-901d-3aab54a7d598-scripts\") pod \"224d1827-61ff-4882-901d-3aab54a7d598\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.208338 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-run\") pod \"224d1827-61ff-4882-901d-3aab54a7d598\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.208398 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-log-ovn\") pod \"224d1827-61ff-4882-901d-3aab54a7d598\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.208457 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-run-ovn\") pod \"224d1827-61ff-4882-901d-3aab54a7d598\" (UID: \"224d1827-61ff-4882-901d-3aab54a7d598\") " Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.209003 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "224d1827-61ff-4882-901d-3aab54a7d598" (UID: "224d1827-61ff-4882-901d-3aab54a7d598"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.210356 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-run" (OuterVolumeSpecName: "var-run") pod "224d1827-61ff-4882-901d-3aab54a7d598" (UID: "224d1827-61ff-4882-901d-3aab54a7d598"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.210399 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "224d1827-61ff-4882-901d-3aab54a7d598" (UID: "224d1827-61ff-4882-901d-3aab54a7d598"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.211233 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/224d1827-61ff-4882-901d-3aab54a7d598-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "224d1827-61ff-4882-901d-3aab54a7d598" (UID: "224d1827-61ff-4882-901d-3aab54a7d598"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.211597 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/224d1827-61ff-4882-901d-3aab54a7d598-scripts" (OuterVolumeSpecName: "scripts") pod "224d1827-61ff-4882-901d-3aab54a7d598" (UID: "224d1827-61ff-4882-901d-3aab54a7d598"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.220996 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/224d1827-61ff-4882-901d-3aab54a7d598-kube-api-access-59bqd" (OuterVolumeSpecName: "kube-api-access-59bqd") pod "224d1827-61ff-4882-901d-3aab54a7d598" (UID: "224d1827-61ff-4882-901d-3aab54a7d598"). InnerVolumeSpecName "kube-api-access-59bqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.310501 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59bqd\" (UniqueName: \"kubernetes.io/projected/224d1827-61ff-4882-901d-3aab54a7d598-kube-api-access-59bqd\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.310558 4806 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/224d1827-61ff-4882-901d-3aab54a7d598-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.310571 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/224d1827-61ff-4882-901d-3aab54a7d598-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.310592 4806 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-run\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.310605 4806 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.310621 4806 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/224d1827-61ff-4882-901d-3aab54a7d598-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.703649 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-tx487-config-7rm8m" event={"ID":"224d1827-61ff-4882-901d-3aab54a7d598","Type":"ContainerDied","Data":"ebfcdc14dc6ff6ee051d19b403ebe417f6eb1035bb485bcaae058fe648c2e06c"} Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.703694 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebfcdc14dc6ff6ee051d19b403ebe417f6eb1035bb485bcaae058fe648c2e06c" Dec 04 04:13:33 crc kubenswrapper[4806]: I1204 04:13:33.703754 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-tx487-config-7rm8m" Dec 04 04:13:34 crc kubenswrapper[4806]: I1204 04:13:34.193878 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-tx487" Dec 04 04:13:34 crc kubenswrapper[4806]: I1204 04:13:34.303670 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-tx487-config-7rm8m"] Dec 04 04:13:34 crc kubenswrapper[4806]: I1204 04:13:34.320173 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-tx487-config-7rm8m"] Dec 04 04:13:34 crc kubenswrapper[4806]: I1204 04:13:34.718698 4806 generic.go:334] "Generic (PLEG): container finished" podID="85f9dc59-e013-464d-b305-77c85cb82f4a" containerID="c3b495d2aebee31942caf47530245a5afa26f0bc262a7c71fa266c7de2ae1a71" exitCode=0 Dec 04 04:13:34 crc kubenswrapper[4806]: I1204 04:13:34.718867 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t7k9p" event={"ID":"85f9dc59-e013-464d-b305-77c85cb82f4a","Type":"ContainerDied","Data":"c3b495d2aebee31942caf47530245a5afa26f0bc262a7c71fa266c7de2ae1a71"} Dec 04 04:13:35 crc kubenswrapper[4806]: I1204 04:13:35.447261 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="224d1827-61ff-4882-901d-3aab54a7d598" path="/var/lib/kubelet/pods/224d1827-61ff-4882-901d-3aab54a7d598/volumes" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.227237 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.304174 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-combined-ca-bundle\") pod \"85f9dc59-e013-464d-b305-77c85cb82f4a\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.304282 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gdpm\" (UniqueName: \"kubernetes.io/projected/85f9dc59-e013-464d-b305-77c85cb82f4a-kube-api-access-5gdpm\") pod \"85f9dc59-e013-464d-b305-77c85cb82f4a\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.304362 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/85f9dc59-e013-464d-b305-77c85cb82f4a-ring-data-devices\") pod \"85f9dc59-e013-464d-b305-77c85cb82f4a\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.304525 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-swiftconf\") pod \"85f9dc59-e013-464d-b305-77c85cb82f4a\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.304580 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85f9dc59-e013-464d-b305-77c85cb82f4a-scripts\") pod \"85f9dc59-e013-464d-b305-77c85cb82f4a\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.304602 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/85f9dc59-e013-464d-b305-77c85cb82f4a-etc-swift\") pod \"85f9dc59-e013-464d-b305-77c85cb82f4a\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.304649 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-dispersionconf\") pod \"85f9dc59-e013-464d-b305-77c85cb82f4a\" (UID: \"85f9dc59-e013-464d-b305-77c85cb82f4a\") " Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.309480 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85f9dc59-e013-464d-b305-77c85cb82f4a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "85f9dc59-e013-464d-b305-77c85cb82f4a" (UID: "85f9dc59-e013-464d-b305-77c85cb82f4a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.311175 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85f9dc59-e013-464d-b305-77c85cb82f4a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "85f9dc59-e013-464d-b305-77c85cb82f4a" (UID: "85f9dc59-e013-464d-b305-77c85cb82f4a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.326307 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85f9dc59-e013-464d-b305-77c85cb82f4a-kube-api-access-5gdpm" (OuterVolumeSpecName: "kube-api-access-5gdpm") pod "85f9dc59-e013-464d-b305-77c85cb82f4a" (UID: "85f9dc59-e013-464d-b305-77c85cb82f4a"). InnerVolumeSpecName "kube-api-access-5gdpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.327218 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85f9dc59-e013-464d-b305-77c85cb82f4a-scripts" (OuterVolumeSpecName: "scripts") pod "85f9dc59-e013-464d-b305-77c85cb82f4a" (UID: "85f9dc59-e013-464d-b305-77c85cb82f4a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.338737 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "85f9dc59-e013-464d-b305-77c85cb82f4a" (UID: "85f9dc59-e013-464d-b305-77c85cb82f4a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.343068 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85f9dc59-e013-464d-b305-77c85cb82f4a" (UID: "85f9dc59-e013-464d-b305-77c85cb82f4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.349598 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "85f9dc59-e013-464d-b305-77c85cb82f4a" (UID: "85f9dc59-e013-464d-b305-77c85cb82f4a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.406741 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gdpm\" (UniqueName: \"kubernetes.io/projected/85f9dc59-e013-464d-b305-77c85cb82f4a-kube-api-access-5gdpm\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.406774 4806 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/85f9dc59-e013-464d-b305-77c85cb82f4a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.406783 4806 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.406793 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/85f9dc59-e013-464d-b305-77c85cb82f4a-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.406800 4806 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/85f9dc59-e013-464d-b305-77c85cb82f4a-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.406808 4806 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.406815 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85f9dc59-e013-464d-b305-77c85cb82f4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.738684 4806 generic.go:334] "Generic (PLEG): container finished" podID="f18b8381-c1f9-4e48-9639-adfc5a406726" containerID="12c5f4cdfff6da63c40c019002d93c0627dc153a33cc9d065f4ac4c3b969019d" exitCode=0 Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.738749 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f18b8381-c1f9-4e48-9639-adfc5a406726","Type":"ContainerDied","Data":"12c5f4cdfff6da63c40c019002d93c0627dc153a33cc9d065f4ac4c3b969019d"} Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.743038 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t7k9p" event={"ID":"85f9dc59-e013-464d-b305-77c85cb82f4a","Type":"ContainerDied","Data":"379cbea57b2a270dca26b81f9a364b5b77567393596cde6e7bcf1c9ec2b56405"} Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.743081 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="379cbea57b2a270dca26b81f9a364b5b77567393596cde6e7bcf1c9ec2b56405" Dec 04 04:13:36 crc kubenswrapper[4806]: I1204 04:13:36.743136 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t7k9p" Dec 04 04:13:47 crc kubenswrapper[4806]: E1204 04:13:47.417876 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 04 04:13:47 crc kubenswrapper[4806]: E1204 04:13:47.419260 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdtwf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-r7mb9_openstack(277e6177-7249-4994-9228-c697d5eecff1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:13:47 crc kubenswrapper[4806]: E1204 04:13:47.422039 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-r7mb9" podUID="277e6177-7249-4994-9228-c697d5eecff1" Dec 04 04:13:47 crc kubenswrapper[4806]: I1204 04:13:47.479556 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:47 crc kubenswrapper[4806]: I1204 04:13:47.484478 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/44220187-6c1e-4755-8420-d629dd59a84a-etc-swift\") pod \"swift-storage-0\" (UID: \"44220187-6c1e-4755-8420-d629dd59a84a\") " pod="openstack/swift-storage-0" Dec 04 04:13:47 crc kubenswrapper[4806]: I1204 04:13:47.705129 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 04 04:13:48 crc kubenswrapper[4806]: I1204 04:13:48.220097 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 04 04:13:48 crc kubenswrapper[4806]: I1204 04:13:48.293461 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f18b8381-c1f9-4e48-9639-adfc5a406726","Type":"ContainerStarted","Data":"818f720223746c8f111996df6c232ef39f5651b1db629200e3810d1486707a6f"} Dec 04 04:13:48 crc kubenswrapper[4806]: I1204 04:13:48.293905 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:13:48 crc kubenswrapper[4806]: I1204 04:13:48.294741 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"44220187-6c1e-4755-8420-d629dd59a84a","Type":"ContainerStarted","Data":"7ec7fd309074c2d8e44caed9e22620319e104031c06e3193dd8204c10aa7fbf2"} Dec 04 04:13:48 crc kubenswrapper[4806]: E1204 04:13:48.295498 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-r7mb9" podUID="277e6177-7249-4994-9228-c697d5eecff1" Dec 04 04:13:48 crc kubenswrapper[4806]: I1204 04:13:48.359536 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371945.495262 podStartE2EDuration="1m31.35951285s" podCreationTimestamp="2025-12-04 04:12:17 +0000 UTC" firstStartedPulling="2025-12-04 04:12:20.447789548 +0000 UTC m=+1055.306302476" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:13:48.32328915 +0000 UTC m=+1143.181802078" watchObservedRunningTime="2025-12-04 04:13:48.35951285 +0000 UTC m=+1143.218025798" Dec 04 04:13:49 crc kubenswrapper[4806]: I1204 04:13:49.694009 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.250250 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-pgg86"] Dec 04 04:13:50 crc kubenswrapper[4806]: E1204 04:13:50.250876 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="224d1827-61ff-4882-901d-3aab54a7d598" containerName="ovn-config" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.250893 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="224d1827-61ff-4882-901d-3aab54a7d598" containerName="ovn-config" Dec 04 04:13:50 crc kubenswrapper[4806]: E1204 04:13:50.250943 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85f9dc59-e013-464d-b305-77c85cb82f4a" containerName="swift-ring-rebalance" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.250950 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="85f9dc59-e013-464d-b305-77c85cb82f4a" containerName="swift-ring-rebalance" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.251111 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="85f9dc59-e013-464d-b305-77c85cb82f4a" containerName="swift-ring-rebalance" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.251134 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="224d1827-61ff-4882-901d-3aab54a7d598" containerName="ovn-config" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.251591 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-pgg86" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.315959 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"44220187-6c1e-4755-8420-d629dd59a84a","Type":"ContainerStarted","Data":"a2c913b731f1024102aedb137f12d2f2ad7de155ca85eeb6b2fd80ae3bfe4087"} Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.316003 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"44220187-6c1e-4755-8420-d629dd59a84a","Type":"ContainerStarted","Data":"9978e8d64bffcf19e6e6a090171fef6553737eb44f6fa91e22c012b326c0741b"} Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.316015 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"44220187-6c1e-4755-8420-d629dd59a84a","Type":"ContainerStarted","Data":"cd83d3b146f4f9faa5dea089cb8b5ad243042321424c06b9a957673a29554ee6"} Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.331537 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-0d3f-account-create-update-lkzhq"] Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.332511 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0d3f-account-create-update-lkzhq" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.348153 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.351531 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-pgg86"] Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.370405 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d576c39-2a71-4977-afab-4d1746f7c38f-operator-scripts\") pod \"barbican-db-create-pgg86\" (UID: \"3d576c39-2a71-4977-afab-4d1746f7c38f\") " pod="openstack/barbican-db-create-pgg86" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.370522 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqvkg\" (UniqueName: \"kubernetes.io/projected/3d576c39-2a71-4977-afab-4d1746f7c38f-kube-api-access-mqvkg\") pod \"barbican-db-create-pgg86\" (UID: \"3d576c39-2a71-4977-afab-4d1746f7c38f\") " pod="openstack/barbican-db-create-pgg86" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.403506 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0d3f-account-create-update-lkzhq"] Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.471809 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqvkg\" (UniqueName: \"kubernetes.io/projected/3d576c39-2a71-4977-afab-4d1746f7c38f-kube-api-access-mqvkg\") pod \"barbican-db-create-pgg86\" (UID: \"3d576c39-2a71-4977-afab-4d1746f7c38f\") " pod="openstack/barbican-db-create-pgg86" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.471865 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be066c54-140f-4727-88eb-864a9d366de5-operator-scripts\") pod \"barbican-0d3f-account-create-update-lkzhq\" (UID: \"be066c54-140f-4727-88eb-864a9d366de5\") " pod="openstack/barbican-0d3f-account-create-update-lkzhq" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.471904 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7n84\" (UniqueName: \"kubernetes.io/projected/be066c54-140f-4727-88eb-864a9d366de5-kube-api-access-q7n84\") pod \"barbican-0d3f-account-create-update-lkzhq\" (UID: \"be066c54-140f-4727-88eb-864a9d366de5\") " pod="openstack/barbican-0d3f-account-create-update-lkzhq" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.471986 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d576c39-2a71-4977-afab-4d1746f7c38f-operator-scripts\") pod \"barbican-db-create-pgg86\" (UID: \"3d576c39-2a71-4977-afab-4d1746f7c38f\") " pod="openstack/barbican-db-create-pgg86" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.472676 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d576c39-2a71-4977-afab-4d1746f7c38f-operator-scripts\") pod \"barbican-db-create-pgg86\" (UID: \"3d576c39-2a71-4977-afab-4d1746f7c38f\") " pod="openstack/barbican-db-create-pgg86" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.488491 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-s5fdn"] Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.489653 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-s5fdn" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.517222 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-s5fdn"] Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.548183 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqvkg\" (UniqueName: \"kubernetes.io/projected/3d576c39-2a71-4977-afab-4d1746f7c38f-kube-api-access-mqvkg\") pod \"barbican-db-create-pgg86\" (UID: \"3d576c39-2a71-4977-afab-4d1746f7c38f\") " pod="openstack/barbican-db-create-pgg86" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.574742 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-pgg86" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.597192 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be066c54-140f-4727-88eb-864a9d366de5-operator-scripts\") pod \"barbican-0d3f-account-create-update-lkzhq\" (UID: \"be066c54-140f-4727-88eb-864a9d366de5\") " pod="openstack/barbican-0d3f-account-create-update-lkzhq" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.597528 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2k9r\" (UniqueName: \"kubernetes.io/projected/7199166b-2234-4b45-b086-3129a62fc2fc-kube-api-access-d2k9r\") pod \"cinder-db-create-s5fdn\" (UID: \"7199166b-2234-4b45-b086-3129a62fc2fc\") " pod="openstack/cinder-db-create-s5fdn" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.597703 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7n84\" (UniqueName: \"kubernetes.io/projected/be066c54-140f-4727-88eb-864a9d366de5-kube-api-access-q7n84\") pod \"barbican-0d3f-account-create-update-lkzhq\" (UID: \"be066c54-140f-4727-88eb-864a9d366de5\") " pod="openstack/barbican-0d3f-account-create-update-lkzhq" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.597826 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7199166b-2234-4b45-b086-3129a62fc2fc-operator-scripts\") pod \"cinder-db-create-s5fdn\" (UID: \"7199166b-2234-4b45-b086-3129a62fc2fc\") " pod="openstack/cinder-db-create-s5fdn" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.599703 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be066c54-140f-4727-88eb-864a9d366de5-operator-scripts\") pod \"barbican-0d3f-account-create-update-lkzhq\" (UID: \"be066c54-140f-4727-88eb-864a9d366de5\") " pod="openstack/barbican-0d3f-account-create-update-lkzhq" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.633276 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7n84\" (UniqueName: \"kubernetes.io/projected/be066c54-140f-4727-88eb-864a9d366de5-kube-api-access-q7n84\") pod \"barbican-0d3f-account-create-update-lkzhq\" (UID: \"be066c54-140f-4727-88eb-864a9d366de5\") " pod="openstack/barbican-0d3f-account-create-update-lkzhq" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.649312 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0d3f-account-create-update-lkzhq" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.667618 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-hbvv2"] Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.669142 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-hbvv2" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.676051 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.676437 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-97vcf" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.683296 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.694496 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-2gm7t"] Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.698247 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.699101 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7199166b-2234-4b45-b086-3129a62fc2fc-operator-scripts\") pod \"cinder-db-create-s5fdn\" (UID: \"7199166b-2234-4b45-b086-3129a62fc2fc\") " pod="openstack/cinder-db-create-s5fdn" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.699387 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2k9r\" (UniqueName: \"kubernetes.io/projected/7199166b-2234-4b45-b086-3129a62fc2fc-kube-api-access-d2k9r\") pod \"cinder-db-create-s5fdn\" (UID: \"7199166b-2234-4b45-b086-3129a62fc2fc\") " pod="openstack/cinder-db-create-s5fdn" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.700442 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7199166b-2234-4b45-b086-3129a62fc2fc-operator-scripts\") pod \"cinder-db-create-s5fdn\" (UID: \"7199166b-2234-4b45-b086-3129a62fc2fc\") " pod="openstack/cinder-db-create-s5fdn" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.701095 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2gm7t" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.727738 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-hbvv2"] Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.734318 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-e2df-account-create-update-ch7v6"] Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.735602 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e2df-account-create-update-ch7v6" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.755002 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.758175 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-2gm7t"] Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.820160 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l4qq\" (UniqueName: \"kubernetes.io/projected/3b10867b-8469-4116-ba55-f6a7cc7b8e01-kube-api-access-9l4qq\") pod \"keystone-db-sync-hbvv2\" (UID: \"3b10867b-8469-4116-ba55-f6a7cc7b8e01\") " pod="openstack/keystone-db-sync-hbvv2" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.820366 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b10867b-8469-4116-ba55-f6a7cc7b8e01-config-data\") pod \"keystone-db-sync-hbvv2\" (UID: \"3b10867b-8469-4116-ba55-f6a7cc7b8e01\") " pod="openstack/keystone-db-sync-hbvv2" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.820488 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b10867b-8469-4116-ba55-f6a7cc7b8e01-combined-ca-bundle\") pod \"keystone-db-sync-hbvv2\" (UID: \"3b10867b-8469-4116-ba55-f6a7cc7b8e01\") " pod="openstack/keystone-db-sync-hbvv2" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.868372 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2k9r\" (UniqueName: \"kubernetes.io/projected/7199166b-2234-4b45-b086-3129a62fc2fc-kube-api-access-d2k9r\") pod \"cinder-db-create-s5fdn\" (UID: \"7199166b-2234-4b45-b086-3129a62fc2fc\") " pod="openstack/cinder-db-create-s5fdn" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.880675 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-e2df-account-create-update-ch7v6"] Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.924118 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b10867b-8469-4116-ba55-f6a7cc7b8e01-config-data\") pod \"keystone-db-sync-hbvv2\" (UID: \"3b10867b-8469-4116-ba55-f6a7cc7b8e01\") " pod="openstack/keystone-db-sync-hbvv2" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.924635 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b10867b-8469-4116-ba55-f6a7cc7b8e01-combined-ca-bundle\") pod \"keystone-db-sync-hbvv2\" (UID: \"3b10867b-8469-4116-ba55-f6a7cc7b8e01\") " pod="openstack/keystone-db-sync-hbvv2" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.924797 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea145431-aaa9-4fee-8485-bf510144f89e-operator-scripts\") pod \"neutron-db-create-2gm7t\" (UID: \"ea145431-aaa9-4fee-8485-bf510144f89e\") " pod="openstack/neutron-db-create-2gm7t" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.924915 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a41b88c9-7bfc-47e8-b612-7129c0821be3-operator-scripts\") pod \"cinder-e2df-account-create-update-ch7v6\" (UID: \"a41b88c9-7bfc-47e8-b612-7129c0821be3\") " pod="openstack/cinder-e2df-account-create-update-ch7v6" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.925075 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbgk2\" (UniqueName: \"kubernetes.io/projected/a41b88c9-7bfc-47e8-b612-7129c0821be3-kube-api-access-rbgk2\") pod \"cinder-e2df-account-create-update-ch7v6\" (UID: \"a41b88c9-7bfc-47e8-b612-7129c0821be3\") " pod="openstack/cinder-e2df-account-create-update-ch7v6" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.925213 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pbg7\" (UniqueName: \"kubernetes.io/projected/ea145431-aaa9-4fee-8485-bf510144f89e-kube-api-access-2pbg7\") pod \"neutron-db-create-2gm7t\" (UID: \"ea145431-aaa9-4fee-8485-bf510144f89e\") " pod="openstack/neutron-db-create-2gm7t" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.925356 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l4qq\" (UniqueName: \"kubernetes.io/projected/3b10867b-8469-4116-ba55-f6a7cc7b8e01-kube-api-access-9l4qq\") pod \"keystone-db-sync-hbvv2\" (UID: \"3b10867b-8469-4116-ba55-f6a7cc7b8e01\") " pod="openstack/keystone-db-sync-hbvv2" Dec 04 04:13:50 crc kubenswrapper[4806]: I1204 04:13:50.936760 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b10867b-8469-4116-ba55-f6a7cc7b8e01-config-data\") pod \"keystone-db-sync-hbvv2\" (UID: \"3b10867b-8469-4116-ba55-f6a7cc7b8e01\") " pod="openstack/keystone-db-sync-hbvv2" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.000615 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b10867b-8469-4116-ba55-f6a7cc7b8e01-combined-ca-bundle\") pod \"keystone-db-sync-hbvv2\" (UID: \"3b10867b-8469-4116-ba55-f6a7cc7b8e01\") " pod="openstack/keystone-db-sync-hbvv2" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.011897 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l4qq\" (UniqueName: \"kubernetes.io/projected/3b10867b-8469-4116-ba55-f6a7cc7b8e01-kube-api-access-9l4qq\") pod \"keystone-db-sync-hbvv2\" (UID: \"3b10867b-8469-4116-ba55-f6a7cc7b8e01\") " pod="openstack/keystone-db-sync-hbvv2" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.017485 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6ba4-account-create-update-nzmnw"] Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.019112 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ba4-account-create-update-nzmnw" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.027075 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pbg7\" (UniqueName: \"kubernetes.io/projected/ea145431-aaa9-4fee-8485-bf510144f89e-kube-api-access-2pbg7\") pod \"neutron-db-create-2gm7t\" (UID: \"ea145431-aaa9-4fee-8485-bf510144f89e\") " pod="openstack/neutron-db-create-2gm7t" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.027237 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea145431-aaa9-4fee-8485-bf510144f89e-operator-scripts\") pod \"neutron-db-create-2gm7t\" (UID: \"ea145431-aaa9-4fee-8485-bf510144f89e\") " pod="openstack/neutron-db-create-2gm7t" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.027262 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a41b88c9-7bfc-47e8-b612-7129c0821be3-operator-scripts\") pod \"cinder-e2df-account-create-update-ch7v6\" (UID: \"a41b88c9-7bfc-47e8-b612-7129c0821be3\") " pod="openstack/cinder-e2df-account-create-update-ch7v6" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.027304 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbgk2\" (UniqueName: \"kubernetes.io/projected/a41b88c9-7bfc-47e8-b612-7129c0821be3-kube-api-access-rbgk2\") pod \"cinder-e2df-account-create-update-ch7v6\" (UID: \"a41b88c9-7bfc-47e8-b612-7129c0821be3\") " pod="openstack/cinder-e2df-account-create-update-ch7v6" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.027905 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.028891 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea145431-aaa9-4fee-8485-bf510144f89e-operator-scripts\") pod \"neutron-db-create-2gm7t\" (UID: \"ea145431-aaa9-4fee-8485-bf510144f89e\") " pod="openstack/neutron-db-create-2gm7t" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.038211 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a41b88c9-7bfc-47e8-b612-7129c0821be3-operator-scripts\") pod \"cinder-e2df-account-create-update-ch7v6\" (UID: \"a41b88c9-7bfc-47e8-b612-7129c0821be3\") " pod="openstack/cinder-e2df-account-create-update-ch7v6" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.048630 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6ba4-account-create-update-nzmnw"] Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.059276 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbgk2\" (UniqueName: \"kubernetes.io/projected/a41b88c9-7bfc-47e8-b612-7129c0821be3-kube-api-access-rbgk2\") pod \"cinder-e2df-account-create-update-ch7v6\" (UID: \"a41b88c9-7bfc-47e8-b612-7129c0821be3\") " pod="openstack/cinder-e2df-account-create-update-ch7v6" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.078582 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pbg7\" (UniqueName: \"kubernetes.io/projected/ea145431-aaa9-4fee-8485-bf510144f89e-kube-api-access-2pbg7\") pod \"neutron-db-create-2gm7t\" (UID: \"ea145431-aaa9-4fee-8485-bf510144f89e\") " pod="openstack/neutron-db-create-2gm7t" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.120213 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-s5fdn" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.133793 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb787acc-3d9f-43ea-a016-0232e3bd1750-operator-scripts\") pod \"neutron-6ba4-account-create-update-nzmnw\" (UID: \"fb787acc-3d9f-43ea-a016-0232e3bd1750\") " pod="openstack/neutron-6ba4-account-create-update-nzmnw" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.133969 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shjr9\" (UniqueName: \"kubernetes.io/projected/fb787acc-3d9f-43ea-a016-0232e3bd1750-kube-api-access-shjr9\") pod \"neutron-6ba4-account-create-update-nzmnw\" (UID: \"fb787acc-3d9f-43ea-a016-0232e3bd1750\") " pod="openstack/neutron-6ba4-account-create-update-nzmnw" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.199841 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-hbvv2" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.242640 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb787acc-3d9f-43ea-a016-0232e3bd1750-operator-scripts\") pod \"neutron-6ba4-account-create-update-nzmnw\" (UID: \"fb787acc-3d9f-43ea-a016-0232e3bd1750\") " pod="openstack/neutron-6ba4-account-create-update-nzmnw" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.243129 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shjr9\" (UniqueName: \"kubernetes.io/projected/fb787acc-3d9f-43ea-a016-0232e3bd1750-kube-api-access-shjr9\") pod \"neutron-6ba4-account-create-update-nzmnw\" (UID: \"fb787acc-3d9f-43ea-a016-0232e3bd1750\") " pod="openstack/neutron-6ba4-account-create-update-nzmnw" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.244602 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb787acc-3d9f-43ea-a016-0232e3bd1750-operator-scripts\") pod \"neutron-6ba4-account-create-update-nzmnw\" (UID: \"fb787acc-3d9f-43ea-a016-0232e3bd1750\") " pod="openstack/neutron-6ba4-account-create-update-nzmnw" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.267904 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2gm7t" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.275764 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e2df-account-create-update-ch7v6" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.280876 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shjr9\" (UniqueName: \"kubernetes.io/projected/fb787acc-3d9f-43ea-a016-0232e3bd1750-kube-api-access-shjr9\") pod \"neutron-6ba4-account-create-update-nzmnw\" (UID: \"fb787acc-3d9f-43ea-a016-0232e3bd1750\") " pod="openstack/neutron-6ba4-account-create-update-nzmnw" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.350015 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ba4-account-create-update-nzmnw" Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.382688 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"44220187-6c1e-4755-8420-d629dd59a84a","Type":"ContainerStarted","Data":"e39ee02dc5b75bf1bbc9b1f1795386a3314cf7990bbcf79633e7ecfb7ff9a63a"} Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.384860 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-pgg86"] Dec 04 04:13:51 crc kubenswrapper[4806]: I1204 04:13:51.782065 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-0d3f-account-create-update-lkzhq"] Dec 04 04:13:52 crc kubenswrapper[4806]: I1204 04:13:52.030243 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-s5fdn"] Dec 04 04:13:52 crc kubenswrapper[4806]: I1204 04:13:52.266748 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-hbvv2"] Dec 04 04:13:52 crc kubenswrapper[4806]: I1204 04:13:52.391936 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-pgg86" event={"ID":"3d576c39-2a71-4977-afab-4d1746f7c38f","Type":"ContainerStarted","Data":"004d8da71d4f9d7ffb858dde4d8f7e94b4be954cbcb1a1d9b5a7453542420111"} Dec 04 04:13:52 crc kubenswrapper[4806]: I1204 04:13:52.391973 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-pgg86" event={"ID":"3d576c39-2a71-4977-afab-4d1746f7c38f","Type":"ContainerStarted","Data":"0c31f5d09809d3aabf9a42d37c117be402acc2be2ba3e00b591e78c68c72b05e"} Dec 04 04:13:52 crc kubenswrapper[4806]: I1204 04:13:52.402187 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0d3f-account-create-update-lkzhq" event={"ID":"be066c54-140f-4727-88eb-864a9d366de5","Type":"ContainerStarted","Data":"e205d21ce56305706301333692cf1d160005ed077a450f8885eafc29f5df3ded"} Dec 04 04:13:52 crc kubenswrapper[4806]: I1204 04:13:52.402223 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0d3f-account-create-update-lkzhq" event={"ID":"be066c54-140f-4727-88eb-864a9d366de5","Type":"ContainerStarted","Data":"628e28c4856e0edddda2fa7fcdc4b3b10019a95f694f40b56150b6c12fc1ef6c"} Dec 04 04:13:52 crc kubenswrapper[4806]: I1204 04:13:52.409788 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-s5fdn" event={"ID":"7199166b-2234-4b45-b086-3129a62fc2fc","Type":"ContainerStarted","Data":"df2248cd6cef99d798274e361bbb050ff8710003fca0f80a0e1ee6e16aa4ae2b"} Dec 04 04:13:52 crc kubenswrapper[4806]: I1204 04:13:52.409831 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-s5fdn" event={"ID":"7199166b-2234-4b45-b086-3129a62fc2fc","Type":"ContainerStarted","Data":"0d1543e0cf59b8812725844069c8fcf9694de96b7b3fae15c357adc986bc07d8"} Dec 04 04:13:52 crc kubenswrapper[4806]: I1204 04:13:52.412704 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-hbvv2" event={"ID":"3b10867b-8469-4116-ba55-f6a7cc7b8e01","Type":"ContainerStarted","Data":"f0475234241fc6bbd038564d83acd853d7f32a8ec0a65f3a3d1fbf9b34799a71"} Dec 04 04:13:52 crc kubenswrapper[4806]: I1204 04:13:52.424530 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-pgg86" podStartSLOduration=2.424512493 podStartE2EDuration="2.424512493s" podCreationTimestamp="2025-12-04 04:13:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:13:52.424283106 +0000 UTC m=+1147.282796034" watchObservedRunningTime="2025-12-04 04:13:52.424512493 +0000 UTC m=+1147.283025421" Dec 04 04:13:52 crc kubenswrapper[4806]: I1204 04:13:52.454813 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-s5fdn" podStartSLOduration=2.454788715 podStartE2EDuration="2.454788715s" podCreationTimestamp="2025-12-04 04:13:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:13:52.446782404 +0000 UTC m=+1147.305295332" watchObservedRunningTime="2025-12-04 04:13:52.454788715 +0000 UTC m=+1147.313301643" Dec 04 04:13:53 crc kubenswrapper[4806]: I1204 04:13:53.097669 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-0d3f-account-create-update-lkzhq" podStartSLOduration=3.097654481 podStartE2EDuration="3.097654481s" podCreationTimestamp="2025-12-04 04:13:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:13:52.473497004 +0000 UTC m=+1147.332009932" watchObservedRunningTime="2025-12-04 04:13:53.097654481 +0000 UTC m=+1147.956167409" Dec 04 04:13:53 crc kubenswrapper[4806]: I1204 04:13:53.106699 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-e2df-account-create-update-ch7v6"] Dec 04 04:13:53 crc kubenswrapper[4806]: I1204 04:13:53.143236 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6ba4-account-create-update-nzmnw"] Dec 04 04:13:53 crc kubenswrapper[4806]: I1204 04:13:53.273455 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-2gm7t"] Dec 04 04:13:53 crc kubenswrapper[4806]: I1204 04:13:53.440951 4806 generic.go:334] "Generic (PLEG): container finished" podID="be066c54-140f-4727-88eb-864a9d366de5" containerID="e205d21ce56305706301333692cf1d160005ed077a450f8885eafc29f5df3ded" exitCode=0 Dec 04 04:13:53 crc kubenswrapper[4806]: I1204 04:13:53.444503 4806 generic.go:334] "Generic (PLEG): container finished" podID="7199166b-2234-4b45-b086-3129a62fc2fc" containerID="df2248cd6cef99d798274e361bbb050ff8710003fca0f80a0e1ee6e16aa4ae2b" exitCode=0 Dec 04 04:13:53 crc kubenswrapper[4806]: I1204 04:13:53.448276 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0d3f-account-create-update-lkzhq" event={"ID":"be066c54-140f-4727-88eb-864a9d366de5","Type":"ContainerDied","Data":"e205d21ce56305706301333692cf1d160005ed077a450f8885eafc29f5df3ded"} Dec 04 04:13:53 crc kubenswrapper[4806]: I1204 04:13:53.448328 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-s5fdn" event={"ID":"7199166b-2234-4b45-b086-3129a62fc2fc","Type":"ContainerDied","Data":"df2248cd6cef99d798274e361bbb050ff8710003fca0f80a0e1ee6e16aa4ae2b"} Dec 04 04:13:53 crc kubenswrapper[4806]: I1204 04:13:53.449669 4806 generic.go:334] "Generic (PLEG): container finished" podID="3d576c39-2a71-4977-afab-4d1746f7c38f" containerID="004d8da71d4f9d7ffb858dde4d8f7e94b4be954cbcb1a1d9b5a7453542420111" exitCode=0 Dec 04 04:13:53 crc kubenswrapper[4806]: I1204 04:13:53.449710 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-pgg86" event={"ID":"3d576c39-2a71-4977-afab-4d1746f7c38f","Type":"ContainerDied","Data":"004d8da71d4f9d7ffb858dde4d8f7e94b4be954cbcb1a1d9b5a7453542420111"} Dec 04 04:13:54 crc kubenswrapper[4806]: I1204 04:13:54.459445 4806 generic.go:334] "Generic (PLEG): container finished" podID="ea145431-aaa9-4fee-8485-bf510144f89e" containerID="0bb85df53d42f9e7a74f60699fb3a94b3aa5b76911eb3437eb2ea9c79eaf6765" exitCode=0 Dec 04 04:13:54 crc kubenswrapper[4806]: I1204 04:13:54.459497 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2gm7t" event={"ID":"ea145431-aaa9-4fee-8485-bf510144f89e","Type":"ContainerDied","Data":"0bb85df53d42f9e7a74f60699fb3a94b3aa5b76911eb3437eb2ea9c79eaf6765"} Dec 04 04:13:54 crc kubenswrapper[4806]: I1204 04:13:54.461208 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2gm7t" event={"ID":"ea145431-aaa9-4fee-8485-bf510144f89e","Type":"ContainerStarted","Data":"0faa848036bd9e3a998781998bfb41a8b626237b8d742456a81161ff59070e15"} Dec 04 04:13:54 crc kubenswrapper[4806]: I1204 04:13:54.463209 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e2df-account-create-update-ch7v6" event={"ID":"a41b88c9-7bfc-47e8-b612-7129c0821be3","Type":"ContainerStarted","Data":"dc3178cb5098590a65febd9a7c27176e1b7a85beeea2f546c10697bffa660e52"} Dec 04 04:13:54 crc kubenswrapper[4806]: I1204 04:13:54.463249 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e2df-account-create-update-ch7v6" event={"ID":"a41b88c9-7bfc-47e8-b612-7129c0821be3","Type":"ContainerStarted","Data":"e72e5cb1653dfd895f308dcb2f6a45d377bae8cf61bfa3a5bf58f646c043c29b"} Dec 04 04:13:54 crc kubenswrapper[4806]: I1204 04:13:54.464853 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ba4-account-create-update-nzmnw" event={"ID":"fb787acc-3d9f-43ea-a016-0232e3bd1750","Type":"ContainerStarted","Data":"40b75f909ea885f4daa09f616dc25bdd594fe94ddb0d92f924e73b20320cd3ac"} Dec 04 04:13:54 crc kubenswrapper[4806]: I1204 04:13:54.464995 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ba4-account-create-update-nzmnw" event={"ID":"fb787acc-3d9f-43ea-a016-0232e3bd1750","Type":"ContainerStarted","Data":"8f30c210d174c4904bcc8629ce86e3dc88cb4de902582dd312094fcdf4af41ee"} Dec 04 04:13:54 crc kubenswrapper[4806]: I1204 04:13:54.468825 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"44220187-6c1e-4755-8420-d629dd59a84a","Type":"ContainerStarted","Data":"ba4b50e5f79eb8971b5cc1e0b9e264a78fd6ede89ec94e41f172b9257800ce28"} Dec 04 04:13:54 crc kubenswrapper[4806]: I1204 04:13:54.468863 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"44220187-6c1e-4755-8420-d629dd59a84a","Type":"ContainerStarted","Data":"658691f74c7c2a61bd54cd1a00421dc849e3c2020a8ff8fee4d5b1d2320ed529"} Dec 04 04:13:54 crc kubenswrapper[4806]: I1204 04:13:54.530963 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6ba4-account-create-update-nzmnw" podStartSLOduration=4.530905354 podStartE2EDuration="4.530905354s" podCreationTimestamp="2025-12-04 04:13:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:13:54.507426136 +0000 UTC m=+1149.365939064" watchObservedRunningTime="2025-12-04 04:13:54.530905354 +0000 UTC m=+1149.389418282" Dec 04 04:13:54 crc kubenswrapper[4806]: I1204 04:13:54.532125 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-e2df-account-create-update-ch7v6" podStartSLOduration=4.532117362 podStartE2EDuration="4.532117362s" podCreationTimestamp="2025-12-04 04:13:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:13:54.522859541 +0000 UTC m=+1149.381372489" watchObservedRunningTime="2025-12-04 04:13:54.532117362 +0000 UTC m=+1149.390630290" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.183670 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0d3f-account-create-update-lkzhq" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.191976 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-s5fdn" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.198442 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-pgg86" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.240897 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqvkg\" (UniqueName: \"kubernetes.io/projected/3d576c39-2a71-4977-afab-4d1746f7c38f-kube-api-access-mqvkg\") pod \"3d576c39-2a71-4977-afab-4d1746f7c38f\" (UID: \"3d576c39-2a71-4977-afab-4d1746f7c38f\") " Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.240959 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2k9r\" (UniqueName: \"kubernetes.io/projected/7199166b-2234-4b45-b086-3129a62fc2fc-kube-api-access-d2k9r\") pod \"7199166b-2234-4b45-b086-3129a62fc2fc\" (UID: \"7199166b-2234-4b45-b086-3129a62fc2fc\") " Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.240984 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d576c39-2a71-4977-afab-4d1746f7c38f-operator-scripts\") pod \"3d576c39-2a71-4977-afab-4d1746f7c38f\" (UID: \"3d576c39-2a71-4977-afab-4d1746f7c38f\") " Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.240998 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7199166b-2234-4b45-b086-3129a62fc2fc-operator-scripts\") pod \"7199166b-2234-4b45-b086-3129a62fc2fc\" (UID: \"7199166b-2234-4b45-b086-3129a62fc2fc\") " Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.241045 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be066c54-140f-4727-88eb-864a9d366de5-operator-scripts\") pod \"be066c54-140f-4727-88eb-864a9d366de5\" (UID: \"be066c54-140f-4727-88eb-864a9d366de5\") " Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.241093 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7n84\" (UniqueName: \"kubernetes.io/projected/be066c54-140f-4727-88eb-864a9d366de5-kube-api-access-q7n84\") pod \"be066c54-140f-4727-88eb-864a9d366de5\" (UID: \"be066c54-140f-4727-88eb-864a9d366de5\") " Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.242674 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7199166b-2234-4b45-b086-3129a62fc2fc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7199166b-2234-4b45-b086-3129a62fc2fc" (UID: "7199166b-2234-4b45-b086-3129a62fc2fc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.242874 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d576c39-2a71-4977-afab-4d1746f7c38f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3d576c39-2a71-4977-afab-4d1746f7c38f" (UID: "3d576c39-2a71-4977-afab-4d1746f7c38f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.243671 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be066c54-140f-4727-88eb-864a9d366de5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "be066c54-140f-4727-88eb-864a9d366de5" (UID: "be066c54-140f-4727-88eb-864a9d366de5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.248495 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d576c39-2a71-4977-afab-4d1746f7c38f-kube-api-access-mqvkg" (OuterVolumeSpecName: "kube-api-access-mqvkg") pod "3d576c39-2a71-4977-afab-4d1746f7c38f" (UID: "3d576c39-2a71-4977-afab-4d1746f7c38f"). InnerVolumeSpecName "kube-api-access-mqvkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.249639 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be066c54-140f-4727-88eb-864a9d366de5-kube-api-access-q7n84" (OuterVolumeSpecName: "kube-api-access-q7n84") pod "be066c54-140f-4727-88eb-864a9d366de5" (UID: "be066c54-140f-4727-88eb-864a9d366de5"). InnerVolumeSpecName "kube-api-access-q7n84". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.250047 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7199166b-2234-4b45-b086-3129a62fc2fc-kube-api-access-d2k9r" (OuterVolumeSpecName: "kube-api-access-d2k9r") pod "7199166b-2234-4b45-b086-3129a62fc2fc" (UID: "7199166b-2234-4b45-b086-3129a62fc2fc"). InnerVolumeSpecName "kube-api-access-d2k9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.346837 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqvkg\" (UniqueName: \"kubernetes.io/projected/3d576c39-2a71-4977-afab-4d1746f7c38f-kube-api-access-mqvkg\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.346877 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2k9r\" (UniqueName: \"kubernetes.io/projected/7199166b-2234-4b45-b086-3129a62fc2fc-kube-api-access-d2k9r\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.346897 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d576c39-2a71-4977-afab-4d1746f7c38f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.346906 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7199166b-2234-4b45-b086-3129a62fc2fc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.346916 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be066c54-140f-4727-88eb-864a9d366de5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.346943 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7n84\" (UniqueName: \"kubernetes.io/projected/be066c54-140f-4727-88eb-864a9d366de5-kube-api-access-q7n84\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.484805 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-s5fdn" event={"ID":"7199166b-2234-4b45-b086-3129a62fc2fc","Type":"ContainerDied","Data":"0d1543e0cf59b8812725844069c8fcf9694de96b7b3fae15c357adc986bc07d8"} Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.484888 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d1543e0cf59b8812725844069c8fcf9694de96b7b3fae15c357adc986bc07d8" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.484837 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-s5fdn" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.494783 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-pgg86" event={"ID":"3d576c39-2a71-4977-afab-4d1746f7c38f","Type":"ContainerDied","Data":"0c31f5d09809d3aabf9a42d37c117be402acc2be2ba3e00b591e78c68c72b05e"} Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.494823 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c31f5d09809d3aabf9a42d37c117be402acc2be2ba3e00b591e78c68c72b05e" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.494912 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-pgg86" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.501667 4806 generic.go:334] "Generic (PLEG): container finished" podID="a41b88c9-7bfc-47e8-b612-7129c0821be3" containerID="dc3178cb5098590a65febd9a7c27176e1b7a85beeea2f546c10697bffa660e52" exitCode=0 Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.501771 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e2df-account-create-update-ch7v6" event={"ID":"a41b88c9-7bfc-47e8-b612-7129c0821be3","Type":"ContainerDied","Data":"dc3178cb5098590a65febd9a7c27176e1b7a85beeea2f546c10697bffa660e52"} Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.505289 4806 generic.go:334] "Generic (PLEG): container finished" podID="fb787acc-3d9f-43ea-a016-0232e3bd1750" containerID="40b75f909ea885f4daa09f616dc25bdd594fe94ddb0d92f924e73b20320cd3ac" exitCode=0 Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.505375 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ba4-account-create-update-nzmnw" event={"ID":"fb787acc-3d9f-43ea-a016-0232e3bd1750","Type":"ContainerDied","Data":"40b75f909ea885f4daa09f616dc25bdd594fe94ddb0d92f924e73b20320cd3ac"} Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.518085 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"44220187-6c1e-4755-8420-d629dd59a84a","Type":"ContainerStarted","Data":"4f2fdbe1afecfb491ec088fa5d7da321cff071ac228d6c900957677ad8727e32"} Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.518299 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"44220187-6c1e-4755-8420-d629dd59a84a","Type":"ContainerStarted","Data":"ad28738dbc395eb4e59a6352c0edfbd6ecae7d2da38df5e20e4901b92bddb9a1"} Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.527002 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-0d3f-account-create-update-lkzhq" event={"ID":"be066c54-140f-4727-88eb-864a9d366de5","Type":"ContainerDied","Data":"628e28c4856e0edddda2fa7fcdc4b3b10019a95f694f40b56150b6c12fc1ef6c"} Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.527059 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="628e28c4856e0edddda2fa7fcdc4b3b10019a95f694f40b56150b6c12fc1ef6c" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.527117 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-0d3f-account-create-update-lkzhq" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.921880 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2gm7t" Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.994208 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea145431-aaa9-4fee-8485-bf510144f89e-operator-scripts\") pod \"ea145431-aaa9-4fee-8485-bf510144f89e\" (UID: \"ea145431-aaa9-4fee-8485-bf510144f89e\") " Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.994416 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pbg7\" (UniqueName: \"kubernetes.io/projected/ea145431-aaa9-4fee-8485-bf510144f89e-kube-api-access-2pbg7\") pod \"ea145431-aaa9-4fee-8485-bf510144f89e\" (UID: \"ea145431-aaa9-4fee-8485-bf510144f89e\") " Dec 04 04:13:55 crc kubenswrapper[4806]: I1204 04:13:55.996054 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea145431-aaa9-4fee-8485-bf510144f89e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ea145431-aaa9-4fee-8485-bf510144f89e" (UID: "ea145431-aaa9-4fee-8485-bf510144f89e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:56 crc kubenswrapper[4806]: I1204 04:13:56.004472 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea145431-aaa9-4fee-8485-bf510144f89e-kube-api-access-2pbg7" (OuterVolumeSpecName: "kube-api-access-2pbg7") pod "ea145431-aaa9-4fee-8485-bf510144f89e" (UID: "ea145431-aaa9-4fee-8485-bf510144f89e"). InnerVolumeSpecName "kube-api-access-2pbg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:56 crc kubenswrapper[4806]: I1204 04:13:56.096191 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pbg7\" (UniqueName: \"kubernetes.io/projected/ea145431-aaa9-4fee-8485-bf510144f89e-kube-api-access-2pbg7\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:56 crc kubenswrapper[4806]: I1204 04:13:56.096222 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea145431-aaa9-4fee-8485-bf510144f89e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:56 crc kubenswrapper[4806]: I1204 04:13:56.537705 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2gm7t" Dec 04 04:13:56 crc kubenswrapper[4806]: I1204 04:13:56.538801 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2gm7t" event={"ID":"ea145431-aaa9-4fee-8485-bf510144f89e","Type":"ContainerDied","Data":"0faa848036bd9e3a998781998bfb41a8b626237b8d742456a81161ff59070e15"} Dec 04 04:13:56 crc kubenswrapper[4806]: I1204 04:13:56.538836 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0faa848036bd9e3a998781998bfb41a8b626237b8d742456a81161ff59070e15" Dec 04 04:13:56 crc kubenswrapper[4806]: I1204 04:13:56.865307 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e2df-account-create-update-ch7v6" Dec 04 04:13:56 crc kubenswrapper[4806]: I1204 04:13:56.876959 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ba4-account-create-update-nzmnw" Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.013691 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb787acc-3d9f-43ea-a016-0232e3bd1750-operator-scripts\") pod \"fb787acc-3d9f-43ea-a016-0232e3bd1750\" (UID: \"fb787acc-3d9f-43ea-a016-0232e3bd1750\") " Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.014016 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbgk2\" (UniqueName: \"kubernetes.io/projected/a41b88c9-7bfc-47e8-b612-7129c0821be3-kube-api-access-rbgk2\") pod \"a41b88c9-7bfc-47e8-b612-7129c0821be3\" (UID: \"a41b88c9-7bfc-47e8-b612-7129c0821be3\") " Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.014085 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shjr9\" (UniqueName: \"kubernetes.io/projected/fb787acc-3d9f-43ea-a016-0232e3bd1750-kube-api-access-shjr9\") pod \"fb787acc-3d9f-43ea-a016-0232e3bd1750\" (UID: \"fb787acc-3d9f-43ea-a016-0232e3bd1750\") " Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.014174 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a41b88c9-7bfc-47e8-b612-7129c0821be3-operator-scripts\") pod \"a41b88c9-7bfc-47e8-b612-7129c0821be3\" (UID: \"a41b88c9-7bfc-47e8-b612-7129c0821be3\") " Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.015372 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a41b88c9-7bfc-47e8-b612-7129c0821be3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a41b88c9-7bfc-47e8-b612-7129c0821be3" (UID: "a41b88c9-7bfc-47e8-b612-7129c0821be3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.015773 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb787acc-3d9f-43ea-a016-0232e3bd1750-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fb787acc-3d9f-43ea-a016-0232e3bd1750" (UID: "fb787acc-3d9f-43ea-a016-0232e3bd1750"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.021634 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb787acc-3d9f-43ea-a016-0232e3bd1750-kube-api-access-shjr9" (OuterVolumeSpecName: "kube-api-access-shjr9") pod "fb787acc-3d9f-43ea-a016-0232e3bd1750" (UID: "fb787acc-3d9f-43ea-a016-0232e3bd1750"). InnerVolumeSpecName "kube-api-access-shjr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.022870 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a41b88c9-7bfc-47e8-b612-7129c0821be3-kube-api-access-rbgk2" (OuterVolumeSpecName: "kube-api-access-rbgk2") pod "a41b88c9-7bfc-47e8-b612-7129c0821be3" (UID: "a41b88c9-7bfc-47e8-b612-7129c0821be3"). InnerVolumeSpecName "kube-api-access-rbgk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.115640 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbgk2\" (UniqueName: \"kubernetes.io/projected/a41b88c9-7bfc-47e8-b612-7129c0821be3-kube-api-access-rbgk2\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.115673 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shjr9\" (UniqueName: \"kubernetes.io/projected/fb787acc-3d9f-43ea-a016-0232e3bd1750-kube-api-access-shjr9\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.115682 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a41b88c9-7bfc-47e8-b612-7129c0821be3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.115691 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb787acc-3d9f-43ea-a016-0232e3bd1750-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.568007 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e2df-account-create-update-ch7v6" event={"ID":"a41b88c9-7bfc-47e8-b612-7129c0821be3","Type":"ContainerDied","Data":"e72e5cb1653dfd895f308dcb2f6a45d377bae8cf61bfa3a5bf58f646c043c29b"} Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.568061 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e72e5cb1653dfd895f308dcb2f6a45d377bae8cf61bfa3a5bf58f646c043c29b" Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.568167 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e2df-account-create-update-ch7v6" Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.575008 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ba4-account-create-update-nzmnw" event={"ID":"fb787acc-3d9f-43ea-a016-0232e3bd1750","Type":"ContainerDied","Data":"8f30c210d174c4904bcc8629ce86e3dc88cb4de902582dd312094fcdf4af41ee"} Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.575076 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f30c210d174c4904bcc8629ce86e3dc88cb4de902582dd312094fcdf4af41ee" Dec 04 04:13:57 crc kubenswrapper[4806]: I1204 04:13:57.575171 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ba4-account-create-update-nzmnw" Dec 04 04:13:59 crc kubenswrapper[4806]: I1204 04:13:59.288121 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:14:01 crc kubenswrapper[4806]: E1204 04:14:01.015543 4806 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.17:37356->38.102.83.17:37121: read tcp 38.102.83.17:37356->38.102.83.17:37121: read: connection reset by peer Dec 04 04:14:03 crc kubenswrapper[4806]: I1204 04:14:03.627657 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-hbvv2" event={"ID":"3b10867b-8469-4116-ba55-f6a7cc7b8e01","Type":"ContainerStarted","Data":"d4415bfd710e982ef75fd954ec910b098cd548cfe8409bdce6c5cd19c6dcb6c7"} Dec 04 04:14:03 crc kubenswrapper[4806]: I1204 04:14:03.631847 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-r7mb9" event={"ID":"277e6177-7249-4994-9228-c697d5eecff1","Type":"ContainerStarted","Data":"867d2725e08df6e3291adffe6875b5963c1e8bc21ec7f2ee49a37f07b7f1a8e0"} Dec 04 04:14:03 crc kubenswrapper[4806]: I1204 04:14:03.640675 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"44220187-6c1e-4755-8420-d629dd59a84a","Type":"ContainerStarted","Data":"59a3df7387b5d2cca21a1cfd5865ad70f0df8eb153093c9df799027a8feb781d"} Dec 04 04:14:03 crc kubenswrapper[4806]: I1204 04:14:03.640882 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"44220187-6c1e-4755-8420-d629dd59a84a","Type":"ContainerStarted","Data":"8405322c2a17484e15c765066e52fcd0e115d9238380b7bca81f1ba3db36284d"} Dec 04 04:14:03 crc kubenswrapper[4806]: I1204 04:14:03.640959 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"44220187-6c1e-4755-8420-d629dd59a84a","Type":"ContainerStarted","Data":"721caaf3982cecbacfa6652c9797706f15d9fbebd50ed48ffeda65f16371d42e"} Dec 04 04:14:03 crc kubenswrapper[4806]: I1204 04:14:03.641020 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"44220187-6c1e-4755-8420-d629dd59a84a","Type":"ContainerStarted","Data":"c4f522a7b06a0645c9aeb825688eecae720d40f6102e119611df2a6f1a445013"} Dec 04 04:14:03 crc kubenswrapper[4806]: I1204 04:14:03.641076 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"44220187-6c1e-4755-8420-d629dd59a84a","Type":"ContainerStarted","Data":"ef94afe4e288f5b46c25f1bc6fea96939291b65cffcb124a4f34e713590e4009"} Dec 04 04:14:03 crc kubenswrapper[4806]: I1204 04:14:03.654108 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-hbvv2" podStartSLOduration=3.308699632 podStartE2EDuration="13.654086989s" podCreationTimestamp="2025-12-04 04:13:50 +0000 UTC" firstStartedPulling="2025-12-04 04:13:52.296362531 +0000 UTC m=+1147.154875459" lastFinishedPulling="2025-12-04 04:14:02.641749888 +0000 UTC m=+1157.500262816" observedRunningTime="2025-12-04 04:14:03.648027948 +0000 UTC m=+1158.506540866" watchObservedRunningTime="2025-12-04 04:14:03.654086989 +0000 UTC m=+1158.512599917" Dec 04 04:14:04 crc kubenswrapper[4806]: I1204 04:14:04.655916 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"44220187-6c1e-4755-8420-d629dd59a84a","Type":"ContainerStarted","Data":"bbaa49fb85a7b8a9177c6dd82c1a78b45c460d6150c8ef6a8219cea5e57ac2c1"} Dec 04 04:14:04 crc kubenswrapper[4806]: I1204 04:14:04.656272 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"44220187-6c1e-4755-8420-d629dd59a84a","Type":"ContainerStarted","Data":"51aca6946fd70320baf5f250b5bbeaf2ce561b15d5582d07ad41cbe43cedcefa"} Dec 04 04:14:04 crc kubenswrapper[4806]: I1204 04:14:04.696869 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.283927166 podStartE2EDuration="50.696841025s" podCreationTimestamp="2025-12-04 04:13:14 +0000 UTC" firstStartedPulling="2025-12-04 04:13:48.228413335 +0000 UTC m=+1143.086926263" lastFinishedPulling="2025-12-04 04:14:02.641327194 +0000 UTC m=+1157.499840122" observedRunningTime="2025-12-04 04:14:04.694729269 +0000 UTC m=+1159.553242197" watchObservedRunningTime="2025-12-04 04:14:04.696841025 +0000 UTC m=+1159.555353953" Dec 04 04:14:04 crc kubenswrapper[4806]: I1204 04:14:04.699764 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-r7mb9" podStartSLOduration=3.384582104 podStartE2EDuration="36.699746497s" podCreationTimestamp="2025-12-04 04:13:28 +0000 UTC" firstStartedPulling="2025-12-04 04:13:29.326906135 +0000 UTC m=+1124.185419063" lastFinishedPulling="2025-12-04 04:14:02.642070528 +0000 UTC m=+1157.500583456" observedRunningTime="2025-12-04 04:14:03.67767049 +0000 UTC m=+1158.536183418" watchObservedRunningTime="2025-12-04 04:14:04.699746497 +0000 UTC m=+1159.558259425" Dec 04 04:14:04 crc kubenswrapper[4806]: I1204 04:14:04.998982 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-xhzfm"] Dec 04 04:14:04 crc kubenswrapper[4806]: E1204 04:14:04.999617 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be066c54-140f-4727-88eb-864a9d366de5" containerName="mariadb-account-create-update" Dec 04 04:14:04 crc kubenswrapper[4806]: I1204 04:14:04.999688 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="be066c54-140f-4727-88eb-864a9d366de5" containerName="mariadb-account-create-update" Dec 04 04:14:04 crc kubenswrapper[4806]: E1204 04:14:04.999749 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb787acc-3d9f-43ea-a016-0232e3bd1750" containerName="mariadb-account-create-update" Dec 04 04:14:04 crc kubenswrapper[4806]: I1204 04:14:04.999799 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb787acc-3d9f-43ea-a016-0232e3bd1750" containerName="mariadb-account-create-update" Dec 04 04:14:04 crc kubenswrapper[4806]: E1204 04:14:04.999865 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a41b88c9-7bfc-47e8-b612-7129c0821be3" containerName="mariadb-account-create-update" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.000080 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a41b88c9-7bfc-47e8-b612-7129c0821be3" containerName="mariadb-account-create-update" Dec 04 04:14:05 crc kubenswrapper[4806]: E1204 04:14:05.000152 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea145431-aaa9-4fee-8485-bf510144f89e" containerName="mariadb-database-create" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.000205 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea145431-aaa9-4fee-8485-bf510144f89e" containerName="mariadb-database-create" Dec 04 04:14:05 crc kubenswrapper[4806]: E1204 04:14:05.000262 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7199166b-2234-4b45-b086-3129a62fc2fc" containerName="mariadb-database-create" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.000311 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7199166b-2234-4b45-b086-3129a62fc2fc" containerName="mariadb-database-create" Dec 04 04:14:05 crc kubenswrapper[4806]: E1204 04:14:05.000366 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d576c39-2a71-4977-afab-4d1746f7c38f" containerName="mariadb-database-create" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.000414 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d576c39-2a71-4977-afab-4d1746f7c38f" containerName="mariadb-database-create" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.000619 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="a41b88c9-7bfc-47e8-b612-7129c0821be3" containerName="mariadb-account-create-update" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.000682 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d576c39-2a71-4977-afab-4d1746f7c38f" containerName="mariadb-database-create" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.000737 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="be066c54-140f-4727-88eb-864a9d366de5" containerName="mariadb-account-create-update" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.000794 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb787acc-3d9f-43ea-a016-0232e3bd1750" containerName="mariadb-account-create-update" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.000851 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="7199166b-2234-4b45-b086-3129a62fc2fc" containerName="mariadb-database-create" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.000908 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea145431-aaa9-4fee-8485-bf510144f89e" containerName="mariadb-database-create" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.001850 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.014999 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.020701 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-xhzfm"] Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.055400 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-dns-svc\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.055704 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-config\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.055789 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.055868 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.055974 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmszt\" (UniqueName: \"kubernetes.io/projected/206f1b10-cd02-4107-b872-b23a06323878-kube-api-access-fmszt\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.056054 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.157617 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-config\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.157689 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.157732 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.157758 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmszt\" (UniqueName: \"kubernetes.io/projected/206f1b10-cd02-4107-b872-b23a06323878-kube-api-access-fmszt\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.157794 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.157886 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-dns-svc\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.159128 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.159144 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.159174 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-dns-svc\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.159367 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-config\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.160107 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.183553 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmszt\" (UniqueName: \"kubernetes.io/projected/206f1b10-cd02-4107-b872-b23a06323878-kube-api-access-fmszt\") pod \"dnsmasq-dns-764c5664d7-xhzfm\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.319626 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:05 crc kubenswrapper[4806]: I1204 04:14:05.809470 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-xhzfm"] Dec 04 04:14:05 crc kubenswrapper[4806]: W1204 04:14:05.814787 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod206f1b10_cd02_4107_b872_b23a06323878.slice/crio-c857fe39423ca140377f4f89716ff27ce4934589352d22fa242a237768658d20 WatchSource:0}: Error finding container c857fe39423ca140377f4f89716ff27ce4934589352d22fa242a237768658d20: Status 404 returned error can't find the container with id c857fe39423ca140377f4f89716ff27ce4934589352d22fa242a237768658d20 Dec 04 04:14:06 crc kubenswrapper[4806]: I1204 04:14:06.675182 4806 generic.go:334] "Generic (PLEG): container finished" podID="206f1b10-cd02-4107-b872-b23a06323878" containerID="2a96eea8352a9f177ece5c2d5f5b62a3967bf285651f0663ecbef10ffbd2dee1" exitCode=0 Dec 04 04:14:06 crc kubenswrapper[4806]: I1204 04:14:06.675273 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" event={"ID":"206f1b10-cd02-4107-b872-b23a06323878","Type":"ContainerDied","Data":"2a96eea8352a9f177ece5c2d5f5b62a3967bf285651f0663ecbef10ffbd2dee1"} Dec 04 04:14:06 crc kubenswrapper[4806]: I1204 04:14:06.675674 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" event={"ID":"206f1b10-cd02-4107-b872-b23a06323878","Type":"ContainerStarted","Data":"c857fe39423ca140377f4f89716ff27ce4934589352d22fa242a237768658d20"} Dec 04 04:14:07 crc kubenswrapper[4806]: I1204 04:14:07.684438 4806 generic.go:334] "Generic (PLEG): container finished" podID="3b10867b-8469-4116-ba55-f6a7cc7b8e01" containerID="d4415bfd710e982ef75fd954ec910b098cd548cfe8409bdce6c5cd19c6dcb6c7" exitCode=0 Dec 04 04:14:07 crc kubenswrapper[4806]: I1204 04:14:07.684534 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-hbvv2" event={"ID":"3b10867b-8469-4116-ba55-f6a7cc7b8e01","Type":"ContainerDied","Data":"d4415bfd710e982ef75fd954ec910b098cd548cfe8409bdce6c5cd19c6dcb6c7"} Dec 04 04:14:07 crc kubenswrapper[4806]: I1204 04:14:07.687168 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" event={"ID":"206f1b10-cd02-4107-b872-b23a06323878","Type":"ContainerStarted","Data":"9c534fe95be1fd426d9a67b57b58451f221e35c13c529d978480ba22e827bd07"} Dec 04 04:14:07 crc kubenswrapper[4806]: I1204 04:14:07.687401 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:07 crc kubenswrapper[4806]: I1204 04:14:07.726632 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" podStartSLOduration=3.7266095679999998 podStartE2EDuration="3.726609568s" podCreationTimestamp="2025-12-04 04:14:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:14:07.721131596 +0000 UTC m=+1162.579644524" watchObservedRunningTime="2025-12-04 04:14:07.726609568 +0000 UTC m=+1162.585122496" Dec 04 04:14:09 crc kubenswrapper[4806]: I1204 04:14:09.024063 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-hbvv2" Dec 04 04:14:09 crc kubenswrapper[4806]: I1204 04:14:09.125230 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b10867b-8469-4116-ba55-f6a7cc7b8e01-config-data\") pod \"3b10867b-8469-4116-ba55-f6a7cc7b8e01\" (UID: \"3b10867b-8469-4116-ba55-f6a7cc7b8e01\") " Dec 04 04:14:09 crc kubenswrapper[4806]: I1204 04:14:09.125412 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b10867b-8469-4116-ba55-f6a7cc7b8e01-combined-ca-bundle\") pod \"3b10867b-8469-4116-ba55-f6a7cc7b8e01\" (UID: \"3b10867b-8469-4116-ba55-f6a7cc7b8e01\") " Dec 04 04:14:09 crc kubenswrapper[4806]: I1204 04:14:09.125452 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9l4qq\" (UniqueName: \"kubernetes.io/projected/3b10867b-8469-4116-ba55-f6a7cc7b8e01-kube-api-access-9l4qq\") pod \"3b10867b-8469-4116-ba55-f6a7cc7b8e01\" (UID: \"3b10867b-8469-4116-ba55-f6a7cc7b8e01\") " Dec 04 04:14:09 crc kubenswrapper[4806]: I1204 04:14:09.135390 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b10867b-8469-4116-ba55-f6a7cc7b8e01-kube-api-access-9l4qq" (OuterVolumeSpecName: "kube-api-access-9l4qq") pod "3b10867b-8469-4116-ba55-f6a7cc7b8e01" (UID: "3b10867b-8469-4116-ba55-f6a7cc7b8e01"). InnerVolumeSpecName "kube-api-access-9l4qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:14:09 crc kubenswrapper[4806]: I1204 04:14:09.156861 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b10867b-8469-4116-ba55-f6a7cc7b8e01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b10867b-8469-4116-ba55-f6a7cc7b8e01" (UID: "3b10867b-8469-4116-ba55-f6a7cc7b8e01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:14:09 crc kubenswrapper[4806]: I1204 04:14:09.182481 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b10867b-8469-4116-ba55-f6a7cc7b8e01-config-data" (OuterVolumeSpecName: "config-data") pod "3b10867b-8469-4116-ba55-f6a7cc7b8e01" (UID: "3b10867b-8469-4116-ba55-f6a7cc7b8e01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:14:09 crc kubenswrapper[4806]: I1204 04:14:09.227501 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b10867b-8469-4116-ba55-f6a7cc7b8e01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:09 crc kubenswrapper[4806]: I1204 04:14:09.228002 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9l4qq\" (UniqueName: \"kubernetes.io/projected/3b10867b-8469-4116-ba55-f6a7cc7b8e01-kube-api-access-9l4qq\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:09 crc kubenswrapper[4806]: I1204 04:14:09.228020 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b10867b-8469-4116-ba55-f6a7cc7b8e01-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:09 crc kubenswrapper[4806]: I1204 04:14:09.712494 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-hbvv2" event={"ID":"3b10867b-8469-4116-ba55-f6a7cc7b8e01","Type":"ContainerDied","Data":"f0475234241fc6bbd038564d83acd853d7f32a8ec0a65f3a3d1fbf9b34799a71"} Dec 04 04:14:09 crc kubenswrapper[4806]: I1204 04:14:09.712538 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0475234241fc6bbd038564d83acd853d7f32a8ec0a65f3a3d1fbf9b34799a71" Dec 04 04:14:09 crc kubenswrapper[4806]: I1204 04:14:09.712558 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-hbvv2" Dec 04 04:14:09 crc kubenswrapper[4806]: I1204 04:14:09.993743 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-xhzfm"] Dec 04 04:14:09 crc kubenswrapper[4806]: I1204 04:14:09.994045 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" podUID="206f1b10-cd02-4107-b872-b23a06323878" containerName="dnsmasq-dns" containerID="cri-o://9c534fe95be1fd426d9a67b57b58451f221e35c13c529d978480ba22e827bd07" gracePeriod=10 Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.024938 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-6ngj2"] Dec 04 04:14:10 crc kubenswrapper[4806]: E1204 04:14:10.030936 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b10867b-8469-4116-ba55-f6a7cc7b8e01" containerName="keystone-db-sync" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.030975 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b10867b-8469-4116-ba55-f6a7cc7b8e01" containerName="keystone-db-sync" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.031173 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b10867b-8469-4116-ba55-f6a7cc7b8e01" containerName="keystone-db-sync" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.031749 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.035896 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.035984 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.036039 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.036250 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.036387 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-97vcf" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.056027 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6ngj2"] Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.136796 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-ddrkw"] Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.144317 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.147736 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-fernet-keys\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.147820 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbfgc\" (UniqueName: \"kubernetes.io/projected/4e56c280-3337-4b25-9455-1b875f1f1a2c-kube-api-access-fbfgc\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.147860 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-config-data\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.147908 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-credential-keys\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.147967 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-combined-ca-bundle\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.147986 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-scripts\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.199119 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-ddrkw"] Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.249339 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-combined-ca-bundle\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.249792 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-scripts\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.249846 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-fernet-keys\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.249887 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbfgc\" (UniqueName: \"kubernetes.io/projected/4e56c280-3337-4b25-9455-1b875f1f1a2c-kube-api-access-fbfgc\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.249915 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-dns-svc\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.249965 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-config-data\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.249999 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.250039 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-credential-keys\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.250057 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.250079 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.250100 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-config\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.250119 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ml9m\" (UniqueName: \"kubernetes.io/projected/835b3b08-b577-445f-9293-6ab047c5b4bc-kube-api-access-6ml9m\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.269786 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-scripts\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.269902 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-fernet-keys\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.270019 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-combined-ca-bundle\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.270538 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-credential-keys\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.270708 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-config-data\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.287890 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbfgc\" (UniqueName: \"kubernetes.io/projected/4e56c280-3337-4b25-9455-1b875f1f1a2c-kube-api-access-fbfgc\") pod \"keystone-bootstrap-6ngj2\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.352295 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-dns-svc\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.352368 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.352432 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.352465 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.352502 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-config\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.352530 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ml9m\" (UniqueName: \"kubernetes.io/projected/835b3b08-b577-445f-9293-6ab047c5b4bc-kube-api-access-6ml9m\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.353975 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-dns-svc\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.354570 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-ovsdbserver-sb\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.355213 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-ovsdbserver-nb\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.355966 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-dns-swift-storage-0\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.356667 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-config\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.357176 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.415882 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ml9m\" (UniqueName: \"kubernetes.io/projected/835b3b08-b577-445f-9293-6ab047c5b4bc-kube-api-access-6ml9m\") pod \"dnsmasq-dns-5959f8865f-ddrkw\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.420760 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-klcjk"] Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.422531 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-klcjk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.454555 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d8614a49-3244-4947-b65b-bc61a8490e30-config\") pod \"neutron-db-sync-klcjk\" (UID: \"d8614a49-3244-4947-b65b-bc61a8490e30\") " pod="openstack/neutron-db-sync-klcjk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.454653 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmfsn\" (UniqueName: \"kubernetes.io/projected/d8614a49-3244-4947-b65b-bc61a8490e30-kube-api-access-mmfsn\") pod \"neutron-db-sync-klcjk\" (UID: \"d8614a49-3244-4947-b65b-bc61a8490e30\") " pod="openstack/neutron-db-sync-klcjk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.454678 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8614a49-3244-4947-b65b-bc61a8490e30-combined-ca-bundle\") pod \"neutron-db-sync-klcjk\" (UID: \"d8614a49-3244-4947-b65b-bc61a8490e30\") " pod="openstack/neutron-db-sync-klcjk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.464632 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-5788c" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.465321 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.465348 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.480259 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-klcjk"] Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.484466 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.558329 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d8614a49-3244-4947-b65b-bc61a8490e30-config\") pod \"neutron-db-sync-klcjk\" (UID: \"d8614a49-3244-4947-b65b-bc61a8490e30\") " pod="openstack/neutron-db-sync-klcjk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.558408 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmfsn\" (UniqueName: \"kubernetes.io/projected/d8614a49-3244-4947-b65b-bc61a8490e30-kube-api-access-mmfsn\") pod \"neutron-db-sync-klcjk\" (UID: \"d8614a49-3244-4947-b65b-bc61a8490e30\") " pod="openstack/neutron-db-sync-klcjk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.571043 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8614a49-3244-4947-b65b-bc61a8490e30-combined-ca-bundle\") pod \"neutron-db-sync-klcjk\" (UID: \"d8614a49-3244-4947-b65b-bc61a8490e30\") " pod="openstack/neutron-db-sync-klcjk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.577227 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d8614a49-3244-4947-b65b-bc61a8490e30-config\") pod \"neutron-db-sync-klcjk\" (UID: \"d8614a49-3244-4947-b65b-bc61a8490e30\") " pod="openstack/neutron-db-sync-klcjk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.599839 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8614a49-3244-4947-b65b-bc61a8490e30-combined-ca-bundle\") pod \"neutron-db-sync-klcjk\" (UID: \"d8614a49-3244-4947-b65b-bc61a8490e30\") " pod="openstack/neutron-db-sync-klcjk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.599963 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-pl5fx"] Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.601518 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pl5fx" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.620811 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-r76nk"] Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.622561 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.630201 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-jjhn5" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.630453 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.630685 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-ctxrq" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.638552 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.638884 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.646299 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7799cdcb77-g8v6s"] Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.661132 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.704635 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-config-data\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.704865 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-scripts\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.704900 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-etc-machine-id\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.707024 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmfsn\" (UniqueName: \"kubernetes.io/projected/d8614a49-3244-4947-b65b-bc61a8490e30-kube-api-access-mmfsn\") pod \"neutron-db-sync-klcjk\" (UID: \"d8614a49-3244-4947-b65b-bc61a8490e30\") " pod="openstack/neutron-db-sync-klcjk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.714660 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.715028 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.715406 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-sxff2" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.715822 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.842499 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtk77\" (UniqueName: \"kubernetes.io/projected/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-kube-api-access-qtk77\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.842898 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-scripts\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.842942 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-db-sync-config-data\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.842970 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-etc-machine-id\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.843013 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e9cfb9-1741-4a4e-a05d-7d666699876c-combined-ca-bundle\") pod \"barbican-db-sync-pl5fx\" (UID: \"b1e9cfb9-1741-4a4e-a05d-7d666699876c\") " pod="openstack/barbican-db-sync-pl5fx" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.843313 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg2hj\" (UniqueName: \"kubernetes.io/projected/b1e9cfb9-1741-4a4e-a05d-7d666699876c-kube-api-access-mg2hj\") pod \"barbican-db-sync-pl5fx\" (UID: \"b1e9cfb9-1741-4a4e-a05d-7d666699876c\") " pod="openstack/barbican-db-sync-pl5fx" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.843377 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-combined-ca-bundle\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.843440 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-config-data\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.843482 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1e9cfb9-1741-4a4e-a05d-7d666699876c-db-sync-config-data\") pod \"barbican-db-sync-pl5fx\" (UID: \"b1e9cfb9-1741-4a4e-a05d-7d666699876c\") " pod="openstack/barbican-db-sync-pl5fx" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.879981 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-scripts\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.880077 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-etc-machine-id\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.880209 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pl5fx"] Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.921546 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-klcjk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.925101 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7799cdcb77-g8v6s"] Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.959056 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82fa1e70-ba16-45e3-8612-8eb8081d7e18-logs\") pod \"horizon-7799cdcb77-g8v6s\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.959121 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/82fa1e70-ba16-45e3-8612-8eb8081d7e18-horizon-secret-key\") pod \"horizon-7799cdcb77-g8v6s\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.959159 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtk77\" (UniqueName: \"kubernetes.io/projected/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-kube-api-access-qtk77\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.959210 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-db-sync-config-data\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.959245 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82fa1e70-ba16-45e3-8612-8eb8081d7e18-scripts\") pod \"horizon-7799cdcb77-g8v6s\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.959272 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e9cfb9-1741-4a4e-a05d-7d666699876c-combined-ca-bundle\") pod \"barbican-db-sync-pl5fx\" (UID: \"b1e9cfb9-1741-4a4e-a05d-7d666699876c\") " pod="openstack/barbican-db-sync-pl5fx" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.959294 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg2hj\" (UniqueName: \"kubernetes.io/projected/b1e9cfb9-1741-4a4e-a05d-7d666699876c-kube-api-access-mg2hj\") pod \"barbican-db-sync-pl5fx\" (UID: \"b1e9cfb9-1741-4a4e-a05d-7d666699876c\") " pod="openstack/barbican-db-sync-pl5fx" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.959327 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sctpp\" (UniqueName: \"kubernetes.io/projected/82fa1e70-ba16-45e3-8612-8eb8081d7e18-kube-api-access-sctpp\") pod \"horizon-7799cdcb77-g8v6s\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.959374 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-combined-ca-bundle\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.959399 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82fa1e70-ba16-45e3-8612-8eb8081d7e18-config-data\") pod \"horizon-7799cdcb77-g8v6s\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.959450 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1e9cfb9-1741-4a4e-a05d-7d666699876c-db-sync-config-data\") pod \"barbican-db-sync-pl5fx\" (UID: \"b1e9cfb9-1741-4a4e-a05d-7d666699876c\") " pod="openstack/barbican-db-sync-pl5fx" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.985424 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-config-data\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.989637 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e9cfb9-1741-4a4e-a05d-7d666699876c-combined-ca-bundle\") pod \"barbican-db-sync-pl5fx\" (UID: \"b1e9cfb9-1741-4a4e-a05d-7d666699876c\") " pod="openstack/barbican-db-sync-pl5fx" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.991696 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1e9cfb9-1741-4a4e-a05d-7d666699876c-db-sync-config-data\") pod \"barbican-db-sync-pl5fx\" (UID: \"b1e9cfb9-1741-4a4e-a05d-7d666699876c\") " pod="openstack/barbican-db-sync-pl5fx" Dec 04 04:14:10 crc kubenswrapper[4806]: I1204 04:14:10.993499 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-db-sync-config-data\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.043791 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-combined-ca-bundle\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.048882 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-r76nk"] Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.066462 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg2hj\" (UniqueName: \"kubernetes.io/projected/b1e9cfb9-1741-4a4e-a05d-7d666699876c-kube-api-access-mg2hj\") pod \"barbican-db-sync-pl5fx\" (UID: \"b1e9cfb9-1741-4a4e-a05d-7d666699876c\") " pod="openstack/barbican-db-sync-pl5fx" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.068468 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtk77\" (UniqueName: \"kubernetes.io/projected/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-kube-api-access-qtk77\") pod \"cinder-db-sync-r76nk\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.073014 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82fa1e70-ba16-45e3-8612-8eb8081d7e18-logs\") pod \"horizon-7799cdcb77-g8v6s\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.073072 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/82fa1e70-ba16-45e3-8612-8eb8081d7e18-horizon-secret-key\") pod \"horizon-7799cdcb77-g8v6s\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.073111 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82fa1e70-ba16-45e3-8612-8eb8081d7e18-scripts\") pod \"horizon-7799cdcb77-g8v6s\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.073153 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sctpp\" (UniqueName: \"kubernetes.io/projected/82fa1e70-ba16-45e3-8612-8eb8081d7e18-kube-api-access-sctpp\") pod \"horizon-7799cdcb77-g8v6s\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.073179 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82fa1e70-ba16-45e3-8612-8eb8081d7e18-config-data\") pod \"horizon-7799cdcb77-g8v6s\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.074363 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82fa1e70-ba16-45e3-8612-8eb8081d7e18-config-data\") pod \"horizon-7799cdcb77-g8v6s\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.074621 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82fa1e70-ba16-45e3-8612-8eb8081d7e18-logs\") pod \"horizon-7799cdcb77-g8v6s\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.082487 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/82fa1e70-ba16-45e3-8612-8eb8081d7e18-horizon-secret-key\") pod \"horizon-7799cdcb77-g8v6s\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.083114 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82fa1e70-ba16-45e3-8612-8eb8081d7e18-scripts\") pod \"horizon-7799cdcb77-g8v6s\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.091184 4806 generic.go:334] "Generic (PLEG): container finished" podID="206f1b10-cd02-4107-b872-b23a06323878" containerID="9c534fe95be1fd426d9a67b57b58451f221e35c13c529d978480ba22e827bd07" exitCode=0 Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.091299 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" event={"ID":"206f1b10-cd02-4107-b872-b23a06323878","Type":"ContainerDied","Data":"9c534fe95be1fd426d9a67b57b58451f221e35c13c529d978480ba22e827bd07"} Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.131556 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sctpp\" (UniqueName: \"kubernetes.io/projected/82fa1e70-ba16-45e3-8612-8eb8081d7e18-kube-api-access-sctpp\") pod \"horizon-7799cdcb77-g8v6s\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.164027 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.166343 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.175356 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.176249 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.212557 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-gpzjd"] Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.213765 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.218356 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.231509 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vvlq5" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.245218 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.255381 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pl5fx" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.275895 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.283909 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-logs\") pod \"placement-db-sync-gpzjd\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.295154 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99g7p\" (UniqueName: \"kubernetes.io/projected/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-kube-api-access-99g7p\") pod \"placement-db-sync-gpzjd\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.295255 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.295297 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwczp\" (UniqueName: \"kubernetes.io/projected/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-kube-api-access-dwczp\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.295337 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-scripts\") pod \"placement-db-sync-gpzjd\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.295462 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-log-httpd\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.295495 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-config-data\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.295593 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.295733 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-combined-ca-bundle\") pod \"placement-db-sync-gpzjd\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.295906 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-run-httpd\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.302294 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-scripts\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.302406 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-config-data\") pod \"placement-db-sync-gpzjd\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.307069 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-ddrkw"] Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.338115 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-gpzjd"] Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.340194 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-r76nk" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.371037 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-sd8q9"] Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.373062 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.398430 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-sd8q9"] Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.404173 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.404706 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-run-httpd\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.404791 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-scripts\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.404818 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-config-data\") pod \"placement-db-sync-gpzjd\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.404868 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-logs\") pod \"placement-db-sync-gpzjd\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.404891 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99g7p\" (UniqueName: \"kubernetes.io/projected/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-kube-api-access-99g7p\") pod \"placement-db-sync-gpzjd\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.405004 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.405032 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwczp\" (UniqueName: \"kubernetes.io/projected/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-kube-api-access-dwczp\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.405056 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-scripts\") pod \"placement-db-sync-gpzjd\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.405218 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-log-httpd\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.405241 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-config-data\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.405293 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.405365 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-combined-ca-bundle\") pod \"placement-db-sync-gpzjd\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.410485 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-run-httpd\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.411240 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-logs\") pod \"placement-db-sync-gpzjd\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.415061 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-log-httpd\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.418840 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.430296 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-config-data\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.431302 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.471391 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-scripts\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.500496 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5f84b64f4f-k26l6"] Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.502605 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwczp\" (UniqueName: \"kubernetes.io/projected/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-kube-api-access-dwczp\") pod \"ceilometer-0\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.503428 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.508250 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-scripts\") pod \"placement-db-sync-gpzjd\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.509322 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-config\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.509344 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.509390 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4nd2\" (UniqueName: \"kubernetes.io/projected/7b424669-446e-4b8a-87b4-0bf6d179ec25-kube-api-access-b4nd2\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.509436 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.509457 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.509576 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.550349 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f84b64f4f-k26l6"] Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.572316 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.586968 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99g7p\" (UniqueName: \"kubernetes.io/projected/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-kube-api-access-99g7p\") pod \"placement-db-sync-gpzjd\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.608068 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-config-data\") pod \"placement-db-sync-gpzjd\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.610957 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd09d5ae-a73d-4563-ad39-544dc269bb61-config-data\") pod \"horizon-5f84b64f4f-k26l6\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.611013 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.611065 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-config\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.611083 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.611103 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd09d5ae-a73d-4563-ad39-544dc269bb61-logs\") pod \"horizon-5f84b64f4f-k26l6\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.611144 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4nd2\" (UniqueName: \"kubernetes.io/projected/7b424669-446e-4b8a-87b4-0bf6d179ec25-kube-api-access-b4nd2\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.611164 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.611184 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.611212 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd09d5ae-a73d-4563-ad39-544dc269bb61-scripts\") pod \"horizon-5f84b64f4f-k26l6\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.611230 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd09d5ae-a73d-4563-ad39-544dc269bb61-horizon-secret-key\") pod \"horizon-5f84b64f4f-k26l6\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.611251 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g52fm\" (UniqueName: \"kubernetes.io/projected/dd09d5ae-a73d-4563-ad39-544dc269bb61-kube-api-access-g52fm\") pod \"horizon-5f84b64f4f-k26l6\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.612061 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-ovsdbserver-nb\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.613318 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-ovsdbserver-sb\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.614106 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-dns-swift-storage-0\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.617872 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-config\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.628019 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-combined-ca-bundle\") pod \"placement-db-sync-gpzjd\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.647312 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-dns-svc\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.696198 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4nd2\" (UniqueName: \"kubernetes.io/projected/7b424669-446e-4b8a-87b4-0bf6d179ec25-kube-api-access-b4nd2\") pod \"dnsmasq-dns-58dd9ff6bc-sd8q9\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.702613 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.712499 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd09d5ae-a73d-4563-ad39-544dc269bb61-scripts\") pod \"horizon-5f84b64f4f-k26l6\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.712557 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd09d5ae-a73d-4563-ad39-544dc269bb61-horizon-secret-key\") pod \"horizon-5f84b64f4f-k26l6\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.712587 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g52fm\" (UniqueName: \"kubernetes.io/projected/dd09d5ae-a73d-4563-ad39-544dc269bb61-kube-api-access-g52fm\") pod \"horizon-5f84b64f4f-k26l6\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.712635 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd09d5ae-a73d-4563-ad39-544dc269bb61-config-data\") pod \"horizon-5f84b64f4f-k26l6\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.712702 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd09d5ae-a73d-4563-ad39-544dc269bb61-logs\") pod \"horizon-5f84b64f4f-k26l6\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.713222 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd09d5ae-a73d-4563-ad39-544dc269bb61-logs\") pod \"horizon-5f84b64f4f-k26l6\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.714602 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd09d5ae-a73d-4563-ad39-544dc269bb61-config-data\") pod \"horizon-5f84b64f4f-k26l6\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.725316 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd09d5ae-a73d-4563-ad39-544dc269bb61-scripts\") pod \"horizon-5f84b64f4f-k26l6\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.759460 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd09d5ae-a73d-4563-ad39-544dc269bb61-horizon-secret-key\") pod \"horizon-5f84b64f4f-k26l6\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.778218 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g52fm\" (UniqueName: \"kubernetes.io/projected/dd09d5ae-a73d-4563-ad39-544dc269bb61-kube-api-access-g52fm\") pod \"horizon-5f84b64f4f-k26l6\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.801845 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.890713 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gpzjd" Dec 04 04:14:11 crc kubenswrapper[4806]: I1204 04:14:11.898308 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6ngj2"] Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.045124 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.127804 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-dns-swift-storage-0\") pod \"206f1b10-cd02-4107-b872-b23a06323878\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.127885 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-ovsdbserver-nb\") pod \"206f1b10-cd02-4107-b872-b23a06323878\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.127912 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-config\") pod \"206f1b10-cd02-4107-b872-b23a06323878\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.128050 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-dns-svc\") pod \"206f1b10-cd02-4107-b872-b23a06323878\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.128119 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmszt\" (UniqueName: \"kubernetes.io/projected/206f1b10-cd02-4107-b872-b23a06323878-kube-api-access-fmszt\") pod \"206f1b10-cd02-4107-b872-b23a06323878\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.128168 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-ovsdbserver-sb\") pod \"206f1b10-cd02-4107-b872-b23a06323878\" (UID: \"206f1b10-cd02-4107-b872-b23a06323878\") " Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.135318 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/206f1b10-cd02-4107-b872-b23a06323878-kube-api-access-fmszt" (OuterVolumeSpecName: "kube-api-access-fmszt") pod "206f1b10-cd02-4107-b872-b23a06323878" (UID: "206f1b10-cd02-4107-b872-b23a06323878"). InnerVolumeSpecName "kube-api-access-fmszt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.154773 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" event={"ID":"206f1b10-cd02-4107-b872-b23a06323878","Type":"ContainerDied","Data":"c857fe39423ca140377f4f89716ff27ce4934589352d22fa242a237768658d20"} Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.155021 4806 scope.go:117] "RemoveContainer" containerID="9c534fe95be1fd426d9a67b57b58451f221e35c13c529d978480ba22e827bd07" Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.155199 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-xhzfm" Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.199564 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6ngj2" event={"ID":"4e56c280-3337-4b25-9455-1b875f1f1a2c","Type":"ContainerStarted","Data":"4dbcdf753c65b4fc466fea7430abba4d52248f63150698f613008cd817c51c9a"} Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.257687 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-ddrkw"] Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.262358 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmszt\" (UniqueName: \"kubernetes.io/projected/206f1b10-cd02-4107-b872-b23a06323878-kube-api-access-fmszt\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.344446 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-config" (OuterVolumeSpecName: "config") pod "206f1b10-cd02-4107-b872-b23a06323878" (UID: "206f1b10-cd02-4107-b872-b23a06323878"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.364622 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.366196 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "206f1b10-cd02-4107-b872-b23a06323878" (UID: "206f1b10-cd02-4107-b872-b23a06323878"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.370295 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "206f1b10-cd02-4107-b872-b23a06323878" (UID: "206f1b10-cd02-4107-b872-b23a06323878"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.391217 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "206f1b10-cd02-4107-b872-b23a06323878" (UID: "206f1b10-cd02-4107-b872-b23a06323878"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.416249 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "206f1b10-cd02-4107-b872-b23a06323878" (UID: "206f1b10-cd02-4107-b872-b23a06323878"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.473535 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.473659 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.473718 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.473769 4806 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/206f1b10-cd02-4107-b872-b23a06323878-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.477380 4806 scope.go:117] "RemoveContainer" containerID="2a96eea8352a9f177ece5c2d5f5b62a3967bf285651f0663ecbef10ffbd2dee1" Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.592221 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-xhzfm"] Dec 04 04:14:12 crc kubenswrapper[4806]: W1204 04:14:12.620463 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8614a49_3244_4947_b65b_bc61a8490e30.slice/crio-a4528ce21e60a8b6e06f1e762311c1685b755e9bd1bb9691d03f282ee9ccda0a WatchSource:0}: Error finding container a4528ce21e60a8b6e06f1e762311c1685b755e9bd1bb9691d03f282ee9ccda0a: Status 404 returned error can't find the container with id a4528ce21e60a8b6e06f1e762311c1685b755e9bd1bb9691d03f282ee9ccda0a Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.631963 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-xhzfm"] Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.670630 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-klcjk"] Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.682063 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-pl5fx"] Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.692008 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-r76nk"] Dec 04 04:14:12 crc kubenswrapper[4806]: I1204 04:14:12.783479 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7799cdcb77-g8v6s"] Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.238033 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-sd8q9"] Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.242780 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-klcjk" event={"ID":"d8614a49-3244-4947-b65b-bc61a8490e30","Type":"ContainerStarted","Data":"311eefa3fc70ca85a9f147992570094cd8822df0a7261165098da44e9736e47f"} Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.242843 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-klcjk" event={"ID":"d8614a49-3244-4947-b65b-bc61a8490e30","Type":"ContainerStarted","Data":"a4528ce21e60a8b6e06f1e762311c1685b755e9bd1bb9691d03f282ee9ccda0a"} Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.266976 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.278933 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-gpzjd"] Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.294034 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-klcjk" podStartSLOduration=3.294004279 podStartE2EDuration="3.294004279s" podCreationTimestamp="2025-12-04 04:14:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:14:13.272384659 +0000 UTC m=+1168.130897587" watchObservedRunningTime="2025-12-04 04:14:13.294004279 +0000 UTC m=+1168.152517207" Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.294110 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-r76nk" event={"ID":"a539706c-9990-4bf7-b393-6fe0bc9f3b0e","Type":"ContainerStarted","Data":"7d5376d5c16c194bfdbd9d0c9de7f1c88c14c0cafe1984a1e6bb4e95461993d4"} Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.296195 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7799cdcb77-g8v6s" event={"ID":"82fa1e70-ba16-45e3-8612-8eb8081d7e18","Type":"ContainerStarted","Data":"315a369f532a44a248a346ff68dfd735c50f84049efb8d1a580134dd2395965d"} Dec 04 04:14:13 crc kubenswrapper[4806]: E1204 04:14:13.331044 4806 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod835b3b08_b577_445f_9293_6ab047c5b4bc.slice/crio-conmon-b1cd593b1f3fb6814aa922e61dff6959382a1ecfd614499a29c7d5854516b76e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod835b3b08_b577_445f_9293_6ab047c5b4bc.slice/crio-b1cd593b1f3fb6814aa922e61dff6959382a1ecfd614499a29c7d5854516b76e.scope\": RecentStats: unable to find data in memory cache]" Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.334465 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6ngj2" event={"ID":"4e56c280-3337-4b25-9455-1b875f1f1a2c","Type":"ContainerStarted","Data":"f8ec359ef9aff652843fea70fa0ad73ef5db4188b5bc4cb5834e8e3d21ba8191"} Dec 04 04:14:13 crc kubenswrapper[4806]: W1204 04:14:13.360456 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd09d5ae_a73d_4563_ad39_544dc269bb61.slice/crio-5f7dfb85dd88f58cde52e582c159742df6193c8f5f9f79b24678059836c6e0dd WatchSource:0}: Error finding container 5f7dfb85dd88f58cde52e582c159742df6193c8f5f9f79b24678059836c6e0dd: Status 404 returned error can't find the container with id 5f7dfb85dd88f58cde52e582c159742df6193c8f5f9f79b24678059836c6e0dd Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.370871 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f84b64f4f-k26l6"] Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.374910 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pl5fx" event={"ID":"b1e9cfb9-1741-4a4e-a05d-7d666699876c","Type":"ContainerStarted","Data":"8fe43f7521c70aca171b052f533bd124a8c5cd5af3d8c2cbbca4a11e5adb520a"} Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.388111 4806 generic.go:334] "Generic (PLEG): container finished" podID="835b3b08-b577-445f-9293-6ab047c5b4bc" containerID="b1cd593b1f3fb6814aa922e61dff6959382a1ecfd614499a29c7d5854516b76e" exitCode=0 Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.388190 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" event={"ID":"835b3b08-b577-445f-9293-6ab047c5b4bc","Type":"ContainerDied","Data":"b1cd593b1f3fb6814aa922e61dff6959382a1ecfd614499a29c7d5854516b76e"} Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.388223 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" event={"ID":"835b3b08-b577-445f-9293-6ab047c5b4bc","Type":"ContainerStarted","Data":"7aa115c5301188dd939892d9870cdb68f8a41ea3a336691a6f32cdbf381d752c"} Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.393376 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-6ngj2" podStartSLOduration=4.391967022 podStartE2EDuration="4.391967022s" podCreationTimestamp="2025-12-04 04:14:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:14:13.372764807 +0000 UTC m=+1168.231277735" watchObservedRunningTime="2025-12-04 04:14:13.391967022 +0000 UTC m=+1168.250479950" Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.490140 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="206f1b10-cd02-4107-b872-b23a06323878" path="/var/lib/kubelet/pods/206f1b10-cd02-4107-b872-b23a06323878/volumes" Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.771250 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f84b64f4f-k26l6"] Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.830876 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-85d594d9c-vgvnw"] Dec 04 04:14:13 crc kubenswrapper[4806]: E1204 04:14:13.846567 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="206f1b10-cd02-4107-b872-b23a06323878" containerName="dnsmasq-dns" Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.846600 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="206f1b10-cd02-4107-b872-b23a06323878" containerName="dnsmasq-dns" Dec 04 04:14:13 crc kubenswrapper[4806]: E1204 04:14:13.846631 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="206f1b10-cd02-4107-b872-b23a06323878" containerName="init" Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.846637 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="206f1b10-cd02-4107-b872-b23a06323878" containerName="init" Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.846805 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="206f1b10-cd02-4107-b872-b23a06323878" containerName="dnsmasq-dns" Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.847825 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.870279 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85d594d9c-vgvnw"] Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.936039 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/287344d1-3065-48ac-926c-008839edfb1d-config-data\") pod \"horizon-85d594d9c-vgvnw\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.936093 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/287344d1-3065-48ac-926c-008839edfb1d-horizon-secret-key\") pod \"horizon-85d594d9c-vgvnw\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.936117 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/287344d1-3065-48ac-926c-008839edfb1d-logs\") pod \"horizon-85d594d9c-vgvnw\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.936324 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/287344d1-3065-48ac-926c-008839edfb1d-scripts\") pod \"horizon-85d594d9c-vgvnw\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:13 crc kubenswrapper[4806]: I1204 04:14:13.936482 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gxv8\" (UniqueName: \"kubernetes.io/projected/287344d1-3065-48ac-926c-008839edfb1d-kube-api-access-6gxv8\") pod \"horizon-85d594d9c-vgvnw\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.012726 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.016641 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.043866 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/287344d1-3065-48ac-926c-008839edfb1d-config-data\") pod \"horizon-85d594d9c-vgvnw\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.043944 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/287344d1-3065-48ac-926c-008839edfb1d-horizon-secret-key\") pod \"horizon-85d594d9c-vgvnw\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.043975 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/287344d1-3065-48ac-926c-008839edfb1d-logs\") pod \"horizon-85d594d9c-vgvnw\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.044038 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/287344d1-3065-48ac-926c-008839edfb1d-scripts\") pod \"horizon-85d594d9c-vgvnw\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.044112 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gxv8\" (UniqueName: \"kubernetes.io/projected/287344d1-3065-48ac-926c-008839edfb1d-kube-api-access-6gxv8\") pod \"horizon-85d594d9c-vgvnw\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.045203 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/287344d1-3065-48ac-926c-008839edfb1d-logs\") pod \"horizon-85d594d9c-vgvnw\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.045377 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/287344d1-3065-48ac-926c-008839edfb1d-config-data\") pod \"horizon-85d594d9c-vgvnw\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.046032 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/287344d1-3065-48ac-926c-008839edfb1d-scripts\") pod \"horizon-85d594d9c-vgvnw\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.064685 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/287344d1-3065-48ac-926c-008839edfb1d-horizon-secret-key\") pod \"horizon-85d594d9c-vgvnw\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.084723 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gxv8\" (UniqueName: \"kubernetes.io/projected/287344d1-3065-48ac-926c-008839edfb1d-kube-api-access-6gxv8\") pod \"horizon-85d594d9c-vgvnw\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.145498 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-ovsdbserver-sb\") pod \"835b3b08-b577-445f-9293-6ab047c5b4bc\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.145652 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-dns-swift-storage-0\") pod \"835b3b08-b577-445f-9293-6ab047c5b4bc\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.145700 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-ovsdbserver-nb\") pod \"835b3b08-b577-445f-9293-6ab047c5b4bc\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.145756 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-config\") pod \"835b3b08-b577-445f-9293-6ab047c5b4bc\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.145835 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-dns-svc\") pod \"835b3b08-b577-445f-9293-6ab047c5b4bc\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.145860 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ml9m\" (UniqueName: \"kubernetes.io/projected/835b3b08-b577-445f-9293-6ab047c5b4bc-kube-api-access-6ml9m\") pod \"835b3b08-b577-445f-9293-6ab047c5b4bc\" (UID: \"835b3b08-b577-445f-9293-6ab047c5b4bc\") " Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.159703 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/835b3b08-b577-445f-9293-6ab047c5b4bc-kube-api-access-6ml9m" (OuterVolumeSpecName: "kube-api-access-6ml9m") pod "835b3b08-b577-445f-9293-6ab047c5b4bc" (UID: "835b3b08-b577-445f-9293-6ab047c5b4bc"). InnerVolumeSpecName "kube-api-access-6ml9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.204365 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "835b3b08-b577-445f-9293-6ab047c5b4bc" (UID: "835b3b08-b577-445f-9293-6ab047c5b4bc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.210246 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.217655 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "835b3b08-b577-445f-9293-6ab047c5b4bc" (UID: "835b3b08-b577-445f-9293-6ab047c5b4bc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.240831 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "835b3b08-b577-445f-9293-6ab047c5b4bc" (UID: "835b3b08-b577-445f-9293-6ab047c5b4bc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.249251 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-config" (OuterVolumeSpecName: "config") pod "835b3b08-b577-445f-9293-6ab047c5b4bc" (UID: "835b3b08-b577-445f-9293-6ab047c5b4bc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.258426 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.258483 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.258495 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.258509 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ml9m\" (UniqueName: \"kubernetes.io/projected/835b3b08-b577-445f-9293-6ab047c5b4bc-kube-api-access-6ml9m\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.258527 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.276286 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "835b3b08-b577-445f-9293-6ab047c5b4bc" (UID: "835b3b08-b577-445f-9293-6ab047c5b4bc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.360575 4806 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/835b3b08-b577-445f-9293-6ab047c5b4bc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.417258 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f84b64f4f-k26l6" event={"ID":"dd09d5ae-a73d-4563-ad39-544dc269bb61","Type":"ContainerStarted","Data":"5f7dfb85dd88f58cde52e582c159742df6193c8f5f9f79b24678059836c6e0dd"} Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.425989 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gpzjd" event={"ID":"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc","Type":"ContainerStarted","Data":"765aa2df37d33eea27b8135b297249c3b1f31f37ef6af6b7731df8de975ca976"} Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.432015 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.432738 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5959f8865f-ddrkw" event={"ID":"835b3b08-b577-445f-9293-6ab047c5b4bc","Type":"ContainerDied","Data":"7aa115c5301188dd939892d9870cdb68f8a41ea3a336691a6f32cdbf381d752c"} Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.432813 4806 scope.go:117] "RemoveContainer" containerID="b1cd593b1f3fb6814aa922e61dff6959382a1ecfd614499a29c7d5854516b76e" Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.443892 4806 generic.go:334] "Generic (PLEG): container finished" podID="7b424669-446e-4b8a-87b4-0bf6d179ec25" containerID="d815f839f60d61f905a16d1665c381684b8103d01d3ac7ded4afcf1d92a7fb33" exitCode=0 Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.443984 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" event={"ID":"7b424669-446e-4b8a-87b4-0bf6d179ec25","Type":"ContainerDied","Data":"d815f839f60d61f905a16d1665c381684b8103d01d3ac7ded4afcf1d92a7fb33"} Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.444011 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" event={"ID":"7b424669-446e-4b8a-87b4-0bf6d179ec25","Type":"ContainerStarted","Data":"5f97ccf1123957c3aa023266caad186493cd223e8874c70d32e64d23aa92d6a5"} Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.464565 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4783b024-15d6-44a8-bf7f-5ae7f52ce45f","Type":"ContainerStarted","Data":"206890fc3d206e0496ba8b02ef7455e9d71e3e90c948cd654e3eb23c4c8c3f19"} Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.707907 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-ddrkw"] Dec 04 04:14:14 crc kubenswrapper[4806]: I1204 04:14:14.758170 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5959f8865f-ddrkw"] Dec 04 04:14:15 crc kubenswrapper[4806]: I1204 04:14:15.098852 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85d594d9c-vgvnw"] Dec 04 04:14:15 crc kubenswrapper[4806]: I1204 04:14:15.447574 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="835b3b08-b577-445f-9293-6ab047c5b4bc" path="/var/lib/kubelet/pods/835b3b08-b577-445f-9293-6ab047c5b4bc/volumes" Dec 04 04:14:15 crc kubenswrapper[4806]: I1204 04:14:15.521545 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" event={"ID":"7b424669-446e-4b8a-87b4-0bf6d179ec25","Type":"ContainerStarted","Data":"70dd03754a513fd15cdc90b7eaacfdf332723a08834d57f7868ddfb770449847"} Dec 04 04:14:15 crc kubenswrapper[4806]: I1204 04:14:15.522485 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:15 crc kubenswrapper[4806]: I1204 04:14:15.542265 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85d594d9c-vgvnw" event={"ID":"287344d1-3065-48ac-926c-008839edfb1d","Type":"ContainerStarted","Data":"cac63031c14102f844dcb67378e2451bccd27b9a256e9d470020c349a8bda034"} Dec 04 04:14:15 crc kubenswrapper[4806]: I1204 04:14:15.668339 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" podStartSLOduration=5.668310962 podStartE2EDuration="5.668310962s" podCreationTimestamp="2025-12-04 04:14:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:14:15.64210767 +0000 UTC m=+1170.500620598" watchObservedRunningTime="2025-12-04 04:14:15.668310962 +0000 UTC m=+1170.526823890" Dec 04 04:14:16 crc kubenswrapper[4806]: I1204 04:14:16.573441 4806 generic.go:334] "Generic (PLEG): container finished" podID="277e6177-7249-4994-9228-c697d5eecff1" containerID="867d2725e08df6e3291adffe6875b5963c1e8bc21ec7f2ee49a37f07b7f1a8e0" exitCode=0 Dec 04 04:14:16 crc kubenswrapper[4806]: I1204 04:14:16.573568 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-r7mb9" event={"ID":"277e6177-7249-4994-9228-c697d5eecff1","Type":"ContainerDied","Data":"867d2725e08df6e3291adffe6875b5963c1e8bc21ec7f2ee49a37f07b7f1a8e0"} Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.292876 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-r7mb9" Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.371112 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-combined-ca-bundle\") pod \"277e6177-7249-4994-9228-c697d5eecff1\" (UID: \"277e6177-7249-4994-9228-c697d5eecff1\") " Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.371504 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdtwf\" (UniqueName: \"kubernetes.io/projected/277e6177-7249-4994-9228-c697d5eecff1-kube-api-access-rdtwf\") pod \"277e6177-7249-4994-9228-c697d5eecff1\" (UID: \"277e6177-7249-4994-9228-c697d5eecff1\") " Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.371588 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-config-data\") pod \"277e6177-7249-4994-9228-c697d5eecff1\" (UID: \"277e6177-7249-4994-9228-c697d5eecff1\") " Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.371688 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-db-sync-config-data\") pod \"277e6177-7249-4994-9228-c697d5eecff1\" (UID: \"277e6177-7249-4994-9228-c697d5eecff1\") " Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.390527 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/277e6177-7249-4994-9228-c697d5eecff1-kube-api-access-rdtwf" (OuterVolumeSpecName: "kube-api-access-rdtwf") pod "277e6177-7249-4994-9228-c697d5eecff1" (UID: "277e6177-7249-4994-9228-c697d5eecff1"). InnerVolumeSpecName "kube-api-access-rdtwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.390644 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "277e6177-7249-4994-9228-c697d5eecff1" (UID: "277e6177-7249-4994-9228-c697d5eecff1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.473958 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdtwf\" (UniqueName: \"kubernetes.io/projected/277e6177-7249-4994-9228-c697d5eecff1-kube-api-access-rdtwf\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.473998 4806 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.514150 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "277e6177-7249-4994-9228-c697d5eecff1" (UID: "277e6177-7249-4994-9228-c697d5eecff1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.527380 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-config-data" (OuterVolumeSpecName: "config-data") pod "277e6177-7249-4994-9228-c697d5eecff1" (UID: "277e6177-7249-4994-9228-c697d5eecff1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.576246 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.576282 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/277e6177-7249-4994-9228-c697d5eecff1-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.605439 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-r7mb9" Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.605503 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-r7mb9" event={"ID":"277e6177-7249-4994-9228-c697d5eecff1","Type":"ContainerDied","Data":"01a2d33ae80f60f8f127060b5c6a09d4bd318cf03118146ee26760773d09dda0"} Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.605610 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01a2d33ae80f60f8f127060b5c6a09d4bd318cf03118146ee26760773d09dda0" Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.630799 4806 generic.go:334] "Generic (PLEG): container finished" podID="4e56c280-3337-4b25-9455-1b875f1f1a2c" containerID="f8ec359ef9aff652843fea70fa0ad73ef5db4188b5bc4cb5834e8e3d21ba8191" exitCode=0 Dec 04 04:14:18 crc kubenswrapper[4806]: I1204 04:14:18.631185 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6ngj2" event={"ID":"4e56c280-3337-4b25-9455-1b875f1f1a2c","Type":"ContainerDied","Data":"f8ec359ef9aff652843fea70fa0ad73ef5db4188b5bc4cb5834e8e3d21ba8191"} Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.132764 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-sd8q9"] Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.136181 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" podUID="7b424669-446e-4b8a-87b4-0bf6d179ec25" containerName="dnsmasq-dns" containerID="cri-o://70dd03754a513fd15cdc90b7eaacfdf332723a08834d57f7868ddfb770449847" gracePeriod=10 Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.180709 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-s94pm"] Dec 04 04:14:19 crc kubenswrapper[4806]: E1204 04:14:19.181114 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="835b3b08-b577-445f-9293-6ab047c5b4bc" containerName="init" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.181126 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="835b3b08-b577-445f-9293-6ab047c5b4bc" containerName="init" Dec 04 04:14:19 crc kubenswrapper[4806]: E1204 04:14:19.181148 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="277e6177-7249-4994-9228-c697d5eecff1" containerName="glance-db-sync" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.181155 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="277e6177-7249-4994-9228-c697d5eecff1" containerName="glance-db-sync" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.181329 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="835b3b08-b577-445f-9293-6ab047c5b4bc" containerName="init" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.181342 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="277e6177-7249-4994-9228-c697d5eecff1" containerName="glance-db-sync" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.190745 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.211269 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-s94pm"] Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.290424 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.290500 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-config\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.290544 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.290611 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.290644 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g8z6\" (UniqueName: \"kubernetes.io/projected/60a725fe-fe27-4d81-826d-d86a501d86ec-kube-api-access-8g8z6\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.290693 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.393337 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.393387 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g8z6\" (UniqueName: \"kubernetes.io/projected/60a725fe-fe27-4d81-826d-d86a501d86ec-kube-api-access-8g8z6\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.393432 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.393468 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.393499 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-config\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.393528 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.399798 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.399797 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.400481 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.400883 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-config\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.401029 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.442978 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g8z6\" (UniqueName: \"kubernetes.io/projected/60a725fe-fe27-4d81-826d-d86a501d86ec-kube-api-access-8g8z6\") pod \"dnsmasq-dns-785d8bcb8c-s94pm\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.541892 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7799cdcb77-g8v6s"] Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.604874 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7bc4f74d8d-xj2xn"] Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.606640 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.609644 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.634315 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bc4f74d8d-xj2xn"] Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.644431 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.693410 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.696635 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.699520 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-zbhs5" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.699696 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.700230 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.700938 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzlm9\" (UniqueName: \"kubernetes.io/projected/2265394b-be59-4797-b54d-369e5a5624e0-kube-api-access-kzlm9\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.701001 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2265394b-be59-4797-b54d-369e5a5624e0-config-data\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.701043 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-horizon-secret-key\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.701085 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-horizon-tls-certs\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.701105 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2265394b-be59-4797-b54d-369e5a5624e0-scripts\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.701119 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-combined-ca-bundle\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.701145 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2265394b-be59-4797-b54d-369e5a5624e0-logs\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.720186 4806 generic.go:334] "Generic (PLEG): container finished" podID="7b424669-446e-4b8a-87b4-0bf6d179ec25" containerID="70dd03754a513fd15cdc90b7eaacfdf332723a08834d57f7868ddfb770449847" exitCode=0 Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.720391 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" event={"ID":"7b424669-446e-4b8a-87b4-0bf6d179ec25","Type":"ContainerDied","Data":"70dd03754a513fd15cdc90b7eaacfdf332723a08834d57f7868ddfb770449847"} Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.741137 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.796003 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.797560 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.801693 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.802720 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-combined-ca-bundle\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.802754 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.802787 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2265394b-be59-4797-b54d-369e5a5624e0-logs\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.802823 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6c3203e-b396-46d8-b74d-24ed86ff2a60-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.802857 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzlm9\" (UniqueName: \"kubernetes.io/projected/2265394b-be59-4797-b54d-369e5a5624e0-kube-api-access-kzlm9\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.802889 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqk9j\" (UniqueName: \"kubernetes.io/projected/e6c3203e-b396-46d8-b74d-24ed86ff2a60-kube-api-access-mqk9j\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.802938 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2265394b-be59-4797-b54d-369e5a5624e0-config-data\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.802956 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.802980 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.803008 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.803025 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-horizon-secret-key\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.803059 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6c3203e-b396-46d8-b74d-24ed86ff2a60-logs\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.803085 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-horizon-tls-certs\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.803127 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2265394b-be59-4797-b54d-369e5a5624e0-scripts\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.803850 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2265394b-be59-4797-b54d-369e5a5624e0-scripts\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.805148 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2265394b-be59-4797-b54d-369e5a5624e0-config-data\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.807729 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2265394b-be59-4797-b54d-369e5a5624e0-logs\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.813450 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-horizon-secret-key\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.817615 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-horizon-tls-certs\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.820456 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-combined-ca-bundle\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.842080 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85d594d9c-vgvnw"] Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.860831 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzlm9\" (UniqueName: \"kubernetes.io/projected/2265394b-be59-4797-b54d-369e5a5624e0-kube-api-access-kzlm9\") pod \"horizon-7bc4f74d8d-xj2xn\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.879506 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.904183 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.904233 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.904253 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.904281 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.904296 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.904315 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-logs\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.904338 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psgkc\" (UniqueName: \"kubernetes.io/projected/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-kube-api-access-psgkc\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.904357 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6c3203e-b396-46d8-b74d-24ed86ff2a60-logs\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.904389 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.904404 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-scripts\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.904440 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6c3203e-b396-46d8-b74d-24ed86ff2a60-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.904459 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.904493 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-config-data\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.904514 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqk9j\" (UniqueName: \"kubernetes.io/projected/e6c3203e-b396-46d8-b74d-24ed86ff2a60-kube-api-access-mqk9j\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.907162 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.916390 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.916720 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6c3203e-b396-46d8-b74d-24ed86ff2a60-logs\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.929453 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6c3203e-b396-46d8-b74d-24ed86ff2a60-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.938810 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.941711 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.949322 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:19 crc kubenswrapper[4806]: I1204 04:14:19.950042 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqk9j\" (UniqueName: \"kubernetes.io/projected/e6c3203e-b396-46d8-b74d-24ed86ff2a60-kube-api-access-mqk9j\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.005753 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.015992 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.008073 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.016199 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.016247 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-logs\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.016313 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psgkc\" (UniqueName: \"kubernetes.io/projected/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-kube-api-access-psgkc\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.016430 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-scripts\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.016626 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.016728 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-config-data\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.006259 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d8bd6c96d-4sqhf"] Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.020740 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.022750 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.023107 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-logs\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.027052 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.028206 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.037175 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-scripts\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.028639 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-config-data\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.061530 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psgkc\" (UniqueName: \"kubernetes.io/projected/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-kube-api-access-psgkc\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.062879 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d8bd6c96d-4sqhf"] Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.118827 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d53864-c3c8-4e40-8791-0bac49ba561e-horizon-tls-certs\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.118896 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d2d53864-c3c8-4e40-8791-0bac49ba561e-config-data\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.118996 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4qs6\" (UniqueName: \"kubernetes.io/projected/d2d53864-c3c8-4e40-8791-0bac49ba561e-kube-api-access-s4qs6\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.119025 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d2d53864-c3c8-4e40-8791-0bac49ba561e-horizon-secret-key\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.119085 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2d53864-c3c8-4e40-8791-0bac49ba561e-logs\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.119132 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d53864-c3c8-4e40-8791-0bac49ba561e-combined-ca-bundle\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.119170 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2d53864-c3c8-4e40-8791-0bac49ba561e-scripts\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.150129 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " pod="openstack/glance-default-external-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.222819 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2d53864-c3c8-4e40-8791-0bac49ba561e-scripts\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.223290 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d53864-c3c8-4e40-8791-0bac49ba561e-horizon-tls-certs\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.223326 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d2d53864-c3c8-4e40-8791-0bac49ba561e-config-data\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.223379 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4qs6\" (UniqueName: \"kubernetes.io/projected/d2d53864-c3c8-4e40-8791-0bac49ba561e-kube-api-access-s4qs6\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.223397 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d2d53864-c3c8-4e40-8791-0bac49ba561e-horizon-secret-key\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.223439 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2d53864-c3c8-4e40-8791-0bac49ba561e-logs\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.223465 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d53864-c3c8-4e40-8791-0bac49ba561e-combined-ca-bundle\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.224228 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2d53864-c3c8-4e40-8791-0bac49ba561e-scripts\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.231392 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d53864-c3c8-4e40-8791-0bac49ba561e-combined-ca-bundle\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.231695 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2d53864-c3c8-4e40-8791-0bac49ba561e-logs\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.232716 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d2d53864-c3c8-4e40-8791-0bac49ba561e-config-data\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.237583 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d53864-c3c8-4e40-8791-0bac49ba561e-horizon-tls-certs\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.246126 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4qs6\" (UniqueName: \"kubernetes.io/projected/d2d53864-c3c8-4e40-8791-0bac49ba561e-kube-api-access-s4qs6\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.246446 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d2d53864-c3c8-4e40-8791-0bac49ba561e-horizon-secret-key\") pod \"horizon-7d8bd6c96d-4sqhf\" (UID: \"d2d53864-c3c8-4e40-8791-0bac49ba561e\") " pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.328308 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 04:14:20 crc kubenswrapper[4806]: I1204 04:14:20.375592 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:14:21 crc kubenswrapper[4806]: I1204 04:14:21.706584 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" podUID="7b424669-446e-4b8a-87b4-0bf6d179ec25" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.139:5353: connect: connection refused" Dec 04 04:14:23 crc kubenswrapper[4806]: I1204 04:14:23.751856 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 04:14:23 crc kubenswrapper[4806]: I1204 04:14:23.810591 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.478566 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.670394 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-credential-keys\") pod \"4e56c280-3337-4b25-9455-1b875f1f1a2c\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.670474 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-scripts\") pod \"4e56c280-3337-4b25-9455-1b875f1f1a2c\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.670536 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-fernet-keys\") pod \"4e56c280-3337-4b25-9455-1b875f1f1a2c\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.670606 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-combined-ca-bundle\") pod \"4e56c280-3337-4b25-9455-1b875f1f1a2c\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.670717 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-config-data\") pod \"4e56c280-3337-4b25-9455-1b875f1f1a2c\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.670740 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbfgc\" (UniqueName: \"kubernetes.io/projected/4e56c280-3337-4b25-9455-1b875f1f1a2c-kube-api-access-fbfgc\") pod \"4e56c280-3337-4b25-9455-1b875f1f1a2c\" (UID: \"4e56c280-3337-4b25-9455-1b875f1f1a2c\") " Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.677468 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-scripts" (OuterVolumeSpecName: "scripts") pod "4e56c280-3337-4b25-9455-1b875f1f1a2c" (UID: "4e56c280-3337-4b25-9455-1b875f1f1a2c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.678169 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4e56c280-3337-4b25-9455-1b875f1f1a2c" (UID: "4e56c280-3337-4b25-9455-1b875f1f1a2c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.679730 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4e56c280-3337-4b25-9455-1b875f1f1a2c" (UID: "4e56c280-3337-4b25-9455-1b875f1f1a2c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.681963 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e56c280-3337-4b25-9455-1b875f1f1a2c-kube-api-access-fbfgc" (OuterVolumeSpecName: "kube-api-access-fbfgc") pod "4e56c280-3337-4b25-9455-1b875f1f1a2c" (UID: "4e56c280-3337-4b25-9455-1b875f1f1a2c"). InnerVolumeSpecName "kube-api-access-fbfgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.707143 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e56c280-3337-4b25-9455-1b875f1f1a2c" (UID: "4e56c280-3337-4b25-9455-1b875f1f1a2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.709067 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-config-data" (OuterVolumeSpecName: "config-data") pod "4e56c280-3337-4b25-9455-1b875f1f1a2c" (UID: "4e56c280-3337-4b25-9455-1b875f1f1a2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.772838 4806 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.772873 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.772883 4806 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.772893 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.772903 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e56c280-3337-4b25-9455-1b875f1f1a2c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.772912 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbfgc\" (UniqueName: \"kubernetes.io/projected/4e56c280-3337-4b25-9455-1b875f1f1a2c-kube-api-access-fbfgc\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.829160 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6ngj2" event={"ID":"4e56c280-3337-4b25-9455-1b875f1f1a2c","Type":"ContainerDied","Data":"4dbcdf753c65b4fc466fea7430abba4d52248f63150698f613008cd817c51c9a"} Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.829196 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4dbcdf753c65b4fc466fea7430abba4d52248f63150698f613008cd817c51c9a" Dec 04 04:14:27 crc kubenswrapper[4806]: I1204 04:14:27.829243 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6ngj2" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.581243 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-6ngj2"] Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.590227 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-6ngj2"] Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.682132 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-sbp9s"] Dec 04 04:14:28 crc kubenswrapper[4806]: E1204 04:14:28.682814 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e56c280-3337-4b25-9455-1b875f1f1a2c" containerName="keystone-bootstrap" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.682830 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e56c280-3337-4b25-9455-1b875f1f1a2c" containerName="keystone-bootstrap" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.682996 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e56c280-3337-4b25-9455-1b875f1f1a2c" containerName="keystone-bootstrap" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.683601 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.689797 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.689957 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.690149 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.696000 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.696149 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-97vcf" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.707545 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sbp9s"] Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.790863 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-fernet-keys\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.790989 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-combined-ca-bundle\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.791150 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-credential-keys\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.791239 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-scripts\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.791302 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96s2n\" (UniqueName: \"kubernetes.io/projected/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-kube-api-access-96s2n\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.791496 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-config-data\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.893168 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-credential-keys\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.893245 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-scripts\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.893293 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96s2n\" (UniqueName: \"kubernetes.io/projected/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-kube-api-access-96s2n\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.893386 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-config-data\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.893445 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-fernet-keys\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.893478 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-combined-ca-bundle\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.900286 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-scripts\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.900590 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-combined-ca-bundle\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.900670 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-credential-keys\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.902360 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-fernet-keys\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.902792 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-config-data\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:28 crc kubenswrapper[4806]: I1204 04:14:28.914447 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96s2n\" (UniqueName: \"kubernetes.io/projected/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-kube-api-access-96s2n\") pod \"keystone-bootstrap-sbp9s\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:29 crc kubenswrapper[4806]: I1204 04:14:29.002908 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:14:29 crc kubenswrapper[4806]: I1204 04:14:29.443974 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e56c280-3337-4b25-9455-1b875f1f1a2c" path="/var/lib/kubelet/pods/4e56c280-3337-4b25-9455-1b875f1f1a2c/volumes" Dec 04 04:14:31 crc kubenswrapper[4806]: I1204 04:14:31.705087 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" podUID="7b424669-446e-4b8a-87b4-0bf6d179ec25" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.139:5353: i/o timeout" Dec 04 04:14:35 crc kubenswrapper[4806]: E1204 04:14:35.032598 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 04 04:14:35 crc kubenswrapper[4806]: E1204 04:14:35.033341 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n66ch86h96h94h595h59ch97h5bh99h5f8h5d9h85h698h57fhbh589h5b8hd8h8bhchbhfh687h77hf4h65h5f8hdch87h665h5dh65dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g52fm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5f84b64f4f-k26l6_openstack(dd09d5ae-a73d-4563-ad39-544dc269bb61): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:14:35 crc kubenswrapper[4806]: E1204 04:14:35.040043 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5f84b64f4f-k26l6" podUID="dd09d5ae-a73d-4563-ad39-544dc269bb61" Dec 04 04:14:35 crc kubenswrapper[4806]: E1204 04:14:35.041116 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 04 04:14:35 crc kubenswrapper[4806]: E1204 04:14:35.041239 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n654hdch576hc9hdfh645h5d7hf6h586h65dhbbh694h58ch567h64h5f4h5bch5cbh656h679h57h556h57bhb4hd9h68h58dh7bh66bh8fh8dh64dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6gxv8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-85d594d9c-vgvnw_openstack(287344d1-3065-48ac-926c-008839edfb1d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:14:35 crc kubenswrapper[4806]: E1204 04:14:35.051425 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 04 04:14:35 crc kubenswrapper[4806]: E1204 04:14:35.051598 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n7bh647h67ch5bbh5f6hfh66h5dbh7fhcdh5hc4h59fh685h559h67hd5h5fch684h5c4h68fh54dh59fhb4h5cdh78h574h555h694h84h5d8h554q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sctpp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7799cdcb77-g8v6s_openstack(82fa1e70-ba16-45e3-8612-8eb8081d7e18): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:14:35 crc kubenswrapper[4806]: E1204 04:14:35.055886 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-85d594d9c-vgvnw" podUID="287344d1-3065-48ac-926c-008839edfb1d" Dec 04 04:14:35 crc kubenswrapper[4806]: E1204 04:14:35.056294 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7799cdcb77-g8v6s" podUID="82fa1e70-ba16-45e3-8612-8eb8081d7e18" Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.175135 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.310016 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-dns-svc\") pod \"7b424669-446e-4b8a-87b4-0bf6d179ec25\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.310267 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4nd2\" (UniqueName: \"kubernetes.io/projected/7b424669-446e-4b8a-87b4-0bf6d179ec25-kube-api-access-b4nd2\") pod \"7b424669-446e-4b8a-87b4-0bf6d179ec25\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.310378 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-ovsdbserver-sb\") pod \"7b424669-446e-4b8a-87b4-0bf6d179ec25\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.310496 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-ovsdbserver-nb\") pod \"7b424669-446e-4b8a-87b4-0bf6d179ec25\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.310595 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-config\") pod \"7b424669-446e-4b8a-87b4-0bf6d179ec25\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.310625 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-dns-swift-storage-0\") pod \"7b424669-446e-4b8a-87b4-0bf6d179ec25\" (UID: \"7b424669-446e-4b8a-87b4-0bf6d179ec25\") " Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.333579 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b424669-446e-4b8a-87b4-0bf6d179ec25-kube-api-access-b4nd2" (OuterVolumeSpecName: "kube-api-access-b4nd2") pod "7b424669-446e-4b8a-87b4-0bf6d179ec25" (UID: "7b424669-446e-4b8a-87b4-0bf6d179ec25"). InnerVolumeSpecName "kube-api-access-b4nd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.362523 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7b424669-446e-4b8a-87b4-0bf6d179ec25" (UID: "7b424669-446e-4b8a-87b4-0bf6d179ec25"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.362554 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7b424669-446e-4b8a-87b4-0bf6d179ec25" (UID: "7b424669-446e-4b8a-87b4-0bf6d179ec25"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.372885 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7b424669-446e-4b8a-87b4-0bf6d179ec25" (UID: "7b424669-446e-4b8a-87b4-0bf6d179ec25"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.385978 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7b424669-446e-4b8a-87b4-0bf6d179ec25" (UID: "7b424669-446e-4b8a-87b4-0bf6d179ec25"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.386491 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-config" (OuterVolumeSpecName: "config") pod "7b424669-446e-4b8a-87b4-0bf6d179ec25" (UID: "7b424669-446e-4b8a-87b4-0bf6d179ec25"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.413054 4806 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.413099 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.413114 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4nd2\" (UniqueName: \"kubernetes.io/projected/7b424669-446e-4b8a-87b4-0bf6d179ec25-kube-api-access-b4nd2\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.413129 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.413141 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.413151 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b424669-446e-4b8a-87b4-0bf6d179ec25-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.901950 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" event={"ID":"7b424669-446e-4b8a-87b4-0bf6d179ec25","Type":"ContainerDied","Data":"5f97ccf1123957c3aa023266caad186493cd223e8874c70d32e64d23aa92d6a5"} Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.902074 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.902237 4806 scope.go:117] "RemoveContainer" containerID="70dd03754a513fd15cdc90b7eaacfdf332723a08834d57f7868ddfb770449847" Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.990993 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-sd8q9"] Dec 04 04:14:35 crc kubenswrapper[4806]: I1204 04:14:35.996040 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58dd9ff6bc-sd8q9"] Dec 04 04:14:36 crc kubenswrapper[4806]: I1204 04:14:36.706429 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-58dd9ff6bc-sd8q9" podUID="7b424669-446e-4b8a-87b4-0bf6d179ec25" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.139:5353: i/o timeout" Dec 04 04:14:37 crc kubenswrapper[4806]: I1204 04:14:37.434257 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b424669-446e-4b8a-87b4-0bf6d179ec25" path="/var/lib/kubelet/pods/7b424669-446e-4b8a-87b4-0bf6d179ec25/volumes" Dec 04 04:14:43 crc kubenswrapper[4806]: I1204 04:14:43.988858 4806 generic.go:334] "Generic (PLEG): container finished" podID="d8614a49-3244-4947-b65b-bc61a8490e30" containerID="311eefa3fc70ca85a9f147992570094cd8822df0a7261165098da44e9736e47f" exitCode=0 Dec 04 04:14:43 crc kubenswrapper[4806]: I1204 04:14:43.988944 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-klcjk" event={"ID":"d8614a49-3244-4947-b65b-bc61a8490e30","Type":"ContainerDied","Data":"311eefa3fc70ca85a9f147992570094cd8822df0a7261165098da44e9736e47f"} Dec 04 04:14:45 crc kubenswrapper[4806]: E1204 04:14:45.123977 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 04 04:14:45 crc kubenswrapper[4806]: E1204 04:14:45.124388 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n656h8ch666h56dh5cch94h59h5b6h699h5c4h6fh696h696h87h5bfh67ch5b5h55ch5bfh66dh5ddh99h74h59ch565h69hb8h6h64dh9bh555hbcq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dwczp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(4783b024-15d6-44a8-bf7f-5ae7f52ce45f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:14:47 crc kubenswrapper[4806]: E1204 04:14:47.425169 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 04 04:14:47 crc kubenswrapper[4806]: E1204 04:14:47.425859 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mg2hj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-pl5fx_openstack(b1e9cfb9-1741-4a4e-a05d-7d666699876c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:14:47 crc kubenswrapper[4806]: E1204 04:14:47.427163 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-pl5fx" podUID="b1e9cfb9-1741-4a4e-a05d-7d666699876c" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.541995 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.553820 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.563066 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.656474 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/82fa1e70-ba16-45e3-8612-8eb8081d7e18-horizon-secret-key\") pod \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.656556 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/287344d1-3065-48ac-926c-008839edfb1d-horizon-secret-key\") pod \"287344d1-3065-48ac-926c-008839edfb1d\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.656583 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82fa1e70-ba16-45e3-8612-8eb8081d7e18-scripts\") pod \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.656615 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/287344d1-3065-48ac-926c-008839edfb1d-scripts\") pod \"287344d1-3065-48ac-926c-008839edfb1d\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.656668 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82fa1e70-ba16-45e3-8612-8eb8081d7e18-config-data\") pod \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.656723 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82fa1e70-ba16-45e3-8612-8eb8081d7e18-logs\") pod \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.656739 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gxv8\" (UniqueName: \"kubernetes.io/projected/287344d1-3065-48ac-926c-008839edfb1d-kube-api-access-6gxv8\") pod \"287344d1-3065-48ac-926c-008839edfb1d\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.656771 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sctpp\" (UniqueName: \"kubernetes.io/projected/82fa1e70-ba16-45e3-8612-8eb8081d7e18-kube-api-access-sctpp\") pod \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\" (UID: \"82fa1e70-ba16-45e3-8612-8eb8081d7e18\") " Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.656851 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/287344d1-3065-48ac-926c-008839edfb1d-logs\") pod \"287344d1-3065-48ac-926c-008839edfb1d\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.656888 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/287344d1-3065-48ac-926c-008839edfb1d-config-data\") pod \"287344d1-3065-48ac-926c-008839edfb1d\" (UID: \"287344d1-3065-48ac-926c-008839edfb1d\") " Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.658133 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/287344d1-3065-48ac-926c-008839edfb1d-config-data" (OuterVolumeSpecName: "config-data") pod "287344d1-3065-48ac-926c-008839edfb1d" (UID: "287344d1-3065-48ac-926c-008839edfb1d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.659497 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82fa1e70-ba16-45e3-8612-8eb8081d7e18-config-data" (OuterVolumeSpecName: "config-data") pod "82fa1e70-ba16-45e3-8612-8eb8081d7e18" (UID: "82fa1e70-ba16-45e3-8612-8eb8081d7e18"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.660609 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82fa1e70-ba16-45e3-8612-8eb8081d7e18-logs" (OuterVolumeSpecName: "logs") pod "82fa1e70-ba16-45e3-8612-8eb8081d7e18" (UID: "82fa1e70-ba16-45e3-8612-8eb8081d7e18"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.661002 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82fa1e70-ba16-45e3-8612-8eb8081d7e18-scripts" (OuterVolumeSpecName: "scripts") pod "82fa1e70-ba16-45e3-8612-8eb8081d7e18" (UID: "82fa1e70-ba16-45e3-8612-8eb8081d7e18"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.661381 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/287344d1-3065-48ac-926c-008839edfb1d-scripts" (OuterVolumeSpecName: "scripts") pod "287344d1-3065-48ac-926c-008839edfb1d" (UID: "287344d1-3065-48ac-926c-008839edfb1d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.661660 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/287344d1-3065-48ac-926c-008839edfb1d-logs" (OuterVolumeSpecName: "logs") pod "287344d1-3065-48ac-926c-008839edfb1d" (UID: "287344d1-3065-48ac-926c-008839edfb1d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.662703 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82fa1e70-ba16-45e3-8612-8eb8081d7e18-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "82fa1e70-ba16-45e3-8612-8eb8081d7e18" (UID: "82fa1e70-ba16-45e3-8612-8eb8081d7e18"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.669452 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82fa1e70-ba16-45e3-8612-8eb8081d7e18-kube-api-access-sctpp" (OuterVolumeSpecName: "kube-api-access-sctpp") pod "82fa1e70-ba16-45e3-8612-8eb8081d7e18" (UID: "82fa1e70-ba16-45e3-8612-8eb8081d7e18"). InnerVolumeSpecName "kube-api-access-sctpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.689474 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/287344d1-3065-48ac-926c-008839edfb1d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "287344d1-3065-48ac-926c-008839edfb1d" (UID: "287344d1-3065-48ac-926c-008839edfb1d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.698746 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/287344d1-3065-48ac-926c-008839edfb1d-kube-api-access-6gxv8" (OuterVolumeSpecName: "kube-api-access-6gxv8") pod "287344d1-3065-48ac-926c-008839edfb1d" (UID: "287344d1-3065-48ac-926c-008839edfb1d"). InnerVolumeSpecName "kube-api-access-6gxv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.759644 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd09d5ae-a73d-4563-ad39-544dc269bb61-horizon-secret-key\") pod \"dd09d5ae-a73d-4563-ad39-544dc269bb61\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.759702 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd09d5ae-a73d-4563-ad39-544dc269bb61-logs\") pod \"dd09d5ae-a73d-4563-ad39-544dc269bb61\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.759771 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd09d5ae-a73d-4563-ad39-544dc269bb61-scripts\") pod \"dd09d5ae-a73d-4563-ad39-544dc269bb61\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.759886 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd09d5ae-a73d-4563-ad39-544dc269bb61-config-data\") pod \"dd09d5ae-a73d-4563-ad39-544dc269bb61\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.759949 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g52fm\" (UniqueName: \"kubernetes.io/projected/dd09d5ae-a73d-4563-ad39-544dc269bb61-kube-api-access-g52fm\") pod \"dd09d5ae-a73d-4563-ad39-544dc269bb61\" (UID: \"dd09d5ae-a73d-4563-ad39-544dc269bb61\") " Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.760288 4806 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/287344d1-3065-48ac-926c-008839edfb1d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.760305 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82fa1e70-ba16-45e3-8612-8eb8081d7e18-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.760314 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/287344d1-3065-48ac-926c-008839edfb1d-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.760322 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82fa1e70-ba16-45e3-8612-8eb8081d7e18-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.760332 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gxv8\" (UniqueName: \"kubernetes.io/projected/287344d1-3065-48ac-926c-008839edfb1d-kube-api-access-6gxv8\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.760340 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82fa1e70-ba16-45e3-8612-8eb8081d7e18-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.760349 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sctpp\" (UniqueName: \"kubernetes.io/projected/82fa1e70-ba16-45e3-8612-8eb8081d7e18-kube-api-access-sctpp\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.760357 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/287344d1-3065-48ac-926c-008839edfb1d-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.760364 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/287344d1-3065-48ac-926c-008839edfb1d-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.760372 4806 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/82fa1e70-ba16-45e3-8612-8eb8081d7e18-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.761191 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd09d5ae-a73d-4563-ad39-544dc269bb61-logs" (OuterVolumeSpecName: "logs") pod "dd09d5ae-a73d-4563-ad39-544dc269bb61" (UID: "dd09d5ae-a73d-4563-ad39-544dc269bb61"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.762139 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd09d5ae-a73d-4563-ad39-544dc269bb61-scripts" (OuterVolumeSpecName: "scripts") pod "dd09d5ae-a73d-4563-ad39-544dc269bb61" (UID: "dd09d5ae-a73d-4563-ad39-544dc269bb61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.762634 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd09d5ae-a73d-4563-ad39-544dc269bb61-config-data" (OuterVolumeSpecName: "config-data") pod "dd09d5ae-a73d-4563-ad39-544dc269bb61" (UID: "dd09d5ae-a73d-4563-ad39-544dc269bb61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.763015 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-s94pm"] Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.766125 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd09d5ae-a73d-4563-ad39-544dc269bb61-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "dd09d5ae-a73d-4563-ad39-544dc269bb61" (UID: "dd09d5ae-a73d-4563-ad39-544dc269bb61"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.766183 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd09d5ae-a73d-4563-ad39-544dc269bb61-kube-api-access-g52fm" (OuterVolumeSpecName: "kube-api-access-g52fm") pod "dd09d5ae-a73d-4563-ad39-544dc269bb61" (UID: "dd09d5ae-a73d-4563-ad39-544dc269bb61"). InnerVolumeSpecName "kube-api-access-g52fm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.864319 4806 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd09d5ae-a73d-4563-ad39-544dc269bb61-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.864358 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd09d5ae-a73d-4563-ad39-544dc269bb61-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.864371 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd09d5ae-a73d-4563-ad39-544dc269bb61-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.864381 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd09d5ae-a73d-4563-ad39-544dc269bb61-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:47 crc kubenswrapper[4806]: I1204 04:14:47.864396 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g52fm\" (UniqueName: \"kubernetes.io/projected/dd09d5ae-a73d-4563-ad39-544dc269bb61-kube-api-access-g52fm\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:48 crc kubenswrapper[4806]: I1204 04:14:48.019671 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7799cdcb77-g8v6s" event={"ID":"82fa1e70-ba16-45e3-8612-8eb8081d7e18","Type":"ContainerDied","Data":"315a369f532a44a248a346ff68dfd735c50f84049efb8d1a580134dd2395965d"} Dec 04 04:14:48 crc kubenswrapper[4806]: I1204 04:14:48.019717 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7799cdcb77-g8v6s" Dec 04 04:14:48 crc kubenswrapper[4806]: I1204 04:14:48.022667 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f84b64f4f-k26l6" event={"ID":"dd09d5ae-a73d-4563-ad39-544dc269bb61","Type":"ContainerDied","Data":"5f7dfb85dd88f58cde52e582c159742df6193c8f5f9f79b24678059836c6e0dd"} Dec 04 04:14:48 crc kubenswrapper[4806]: I1204 04:14:48.022749 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f84b64f4f-k26l6" Dec 04 04:14:48 crc kubenswrapper[4806]: I1204 04:14:48.025705 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85d594d9c-vgvnw" event={"ID":"287344d1-3065-48ac-926c-008839edfb1d","Type":"ContainerDied","Data":"cac63031c14102f844dcb67378e2451bccd27b9a256e9d470020c349a8bda034"} Dec 04 04:14:48 crc kubenswrapper[4806]: I1204 04:14:48.025759 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85d594d9c-vgvnw" Dec 04 04:14:48 crc kubenswrapper[4806]: E1204 04:14:48.027176 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-pl5fx" podUID="b1e9cfb9-1741-4a4e-a05d-7d666699876c" Dec 04 04:14:48 crc kubenswrapper[4806]: I1204 04:14:48.092147 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7799cdcb77-g8v6s"] Dec 04 04:14:48 crc kubenswrapper[4806]: I1204 04:14:48.109217 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7799cdcb77-g8v6s"] Dec 04 04:14:48 crc kubenswrapper[4806]: I1204 04:14:48.138499 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f84b64f4f-k26l6"] Dec 04 04:14:48 crc kubenswrapper[4806]: I1204 04:14:48.210398 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5f84b64f4f-k26l6"] Dec 04 04:14:48 crc kubenswrapper[4806]: I1204 04:14:48.240237 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85d594d9c-vgvnw"] Dec 04 04:14:48 crc kubenswrapper[4806]: I1204 04:14:48.247404 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-85d594d9c-vgvnw"] Dec 04 04:14:49 crc kubenswrapper[4806]: I1204 04:14:49.432302 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="287344d1-3065-48ac-926c-008839edfb1d" path="/var/lib/kubelet/pods/287344d1-3065-48ac-926c-008839edfb1d/volumes" Dec 04 04:14:49 crc kubenswrapper[4806]: I1204 04:14:49.433070 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82fa1e70-ba16-45e3-8612-8eb8081d7e18" path="/var/lib/kubelet/pods/82fa1e70-ba16-45e3-8612-8eb8081d7e18/volumes" Dec 04 04:14:49 crc kubenswrapper[4806]: I1204 04:14:49.433514 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd09d5ae-a73d-4563-ad39-544dc269bb61" path="/var/lib/kubelet/pods/dd09d5ae-a73d-4563-ad39-544dc269bb61/volumes" Dec 04 04:14:51 crc kubenswrapper[4806]: E1204 04:14:51.102305 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 04 04:14:51 crc kubenswrapper[4806]: E1204 04:14:51.102475 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qtk77,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-r76nk_openstack(a539706c-9990-4bf7-b393-6fe0bc9f3b0e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:14:51 crc kubenswrapper[4806]: E1204 04:14:51.105788 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-r76nk" podUID="a539706c-9990-4bf7-b393-6fe0bc9f3b0e" Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.123033 4806 scope.go:117] "RemoveContainer" containerID="d815f839f60d61f905a16d1665c381684b8103d01d3ac7ded4afcf1d92a7fb33" Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.331943 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-klcjk" Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.439236 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d8614a49-3244-4947-b65b-bc61a8490e30-config\") pod \"d8614a49-3244-4947-b65b-bc61a8490e30\" (UID: \"d8614a49-3244-4947-b65b-bc61a8490e30\") " Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.439393 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8614a49-3244-4947-b65b-bc61a8490e30-combined-ca-bundle\") pod \"d8614a49-3244-4947-b65b-bc61a8490e30\" (UID: \"d8614a49-3244-4947-b65b-bc61a8490e30\") " Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.439424 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmfsn\" (UniqueName: \"kubernetes.io/projected/d8614a49-3244-4947-b65b-bc61a8490e30-kube-api-access-mmfsn\") pod \"d8614a49-3244-4947-b65b-bc61a8490e30\" (UID: \"d8614a49-3244-4947-b65b-bc61a8490e30\") " Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.446533 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8614a49-3244-4947-b65b-bc61a8490e30-kube-api-access-mmfsn" (OuterVolumeSpecName: "kube-api-access-mmfsn") pod "d8614a49-3244-4947-b65b-bc61a8490e30" (UID: "d8614a49-3244-4947-b65b-bc61a8490e30"). InnerVolumeSpecName "kube-api-access-mmfsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.464315 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8614a49-3244-4947-b65b-bc61a8490e30-config" (OuterVolumeSpecName: "config") pod "d8614a49-3244-4947-b65b-bc61a8490e30" (UID: "d8614a49-3244-4947-b65b-bc61a8490e30"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.465940 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8614a49-3244-4947-b65b-bc61a8490e30-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8614a49-3244-4947-b65b-bc61a8490e30" (UID: "d8614a49-3244-4947-b65b-bc61a8490e30"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.543200 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d8614a49-3244-4947-b65b-bc61a8490e30-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.543257 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8614a49-3244-4947-b65b-bc61a8490e30-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.543275 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmfsn\" (UniqueName: \"kubernetes.io/projected/d8614a49-3244-4947-b65b-bc61a8490e30-kube-api-access-mmfsn\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.699479 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d8bd6c96d-4sqhf"] Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.709473 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bc4f74d8d-xj2xn"] Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.773113 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 04:14:51 crc kubenswrapper[4806]: W1204 04:14:51.777459 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2d53864_c3c8_4e40_8791_0bac49ba561e.slice/crio-42255d71f1b7a5697de2e235f43a385c614250b14520f0b8f1caf66b56524513 WatchSource:0}: Error finding container 42255d71f1b7a5697de2e235f43a385c614250b14520f0b8f1caf66b56524513: Status 404 returned error can't find the container with id 42255d71f1b7a5697de2e235f43a385c614250b14520f0b8f1caf66b56524513 Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.883136 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sbp9s"] Dec 04 04:14:51 crc kubenswrapper[4806]: W1204 04:14:51.910235 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbabd629a_b0d5_4ba9_9e01_d36f6bc9ac36.slice/crio-7c7df44aa0a44a3869e0106629043e8211b2d5a842f3852dac77f017def87eb7 WatchSource:0}: Error finding container 7c7df44aa0a44a3869e0106629043e8211b2d5a842f3852dac77f017def87eb7: Status 404 returned error can't find the container with id 7c7df44aa0a44a3869e0106629043e8211b2d5a842f3852dac77f017def87eb7 Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.927234 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 04:14:51 crc kubenswrapper[4806]: I1204 04:14:51.985359 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.062764 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sbp9s" event={"ID":"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36","Type":"ContainerStarted","Data":"7c7df44aa0a44a3869e0106629043e8211b2d5a842f3852dac77f017def87eb7"} Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.066059 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-klcjk" event={"ID":"d8614a49-3244-4947-b65b-bc61a8490e30","Type":"ContainerDied","Data":"a4528ce21e60a8b6e06f1e762311c1685b755e9bd1bb9691d03f282ee9ccda0a"} Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.066082 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4528ce21e60a8b6e06f1e762311c1685b755e9bd1bb9691d03f282ee9ccda0a" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.066253 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-klcjk" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.070310 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc4f74d8d-xj2xn" event={"ID":"2265394b-be59-4797-b54d-369e5a5624e0","Type":"ContainerStarted","Data":"b961042d0e3bd0d5b615befeef373210308ebe50776af5b705d242c124924f6f"} Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.076679 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e6c3203e-b396-46d8-b74d-24ed86ff2a60","Type":"ContainerStarted","Data":"b5d23decab701a5746d790c14f8b36bc1c010e37ba85829fe55907304998e58b"} Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.082578 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" event={"ID":"60a725fe-fe27-4d81-826d-d86a501d86ec","Type":"ContainerStarted","Data":"3bc188417bac387dce862e2208d528d50a0a495301490586f28ed4127888b2c9"} Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.082637 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" event={"ID":"60a725fe-fe27-4d81-826d-d86a501d86ec","Type":"ContainerStarted","Data":"71bd9b6333c48cf446072af8d54a502b65602d999c18b86ecfd75bbaa6507815"} Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.088769 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d8bd6c96d-4sqhf" event={"ID":"d2d53864-c3c8-4e40-8791-0bac49ba561e","Type":"ContainerStarted","Data":"42255d71f1b7a5697de2e235f43a385c614250b14520f0b8f1caf66b56524513"} Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.091514 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6","Type":"ContainerStarted","Data":"f783b666c8119ac094f3c24bca0cb7252cb1ac304bc0746160703daa8a581f45"} Dec 04 04:14:52 crc kubenswrapper[4806]: E1204 04:14:52.092209 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-r76nk" podUID="a539706c-9990-4bf7-b393-6fe0bc9f3b0e" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.608671 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-s94pm"] Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.729471 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-wbj9t"] Dec 04 04:14:52 crc kubenswrapper[4806]: E1204 04:14:52.730075 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b424669-446e-4b8a-87b4-0bf6d179ec25" containerName="dnsmasq-dns" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.730089 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b424669-446e-4b8a-87b4-0bf6d179ec25" containerName="dnsmasq-dns" Dec 04 04:14:52 crc kubenswrapper[4806]: E1204 04:14:52.730105 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8614a49-3244-4947-b65b-bc61a8490e30" containerName="neutron-db-sync" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.730111 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8614a49-3244-4947-b65b-bc61a8490e30" containerName="neutron-db-sync" Dec 04 04:14:52 crc kubenswrapper[4806]: E1204 04:14:52.730124 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b424669-446e-4b8a-87b4-0bf6d179ec25" containerName="init" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.730131 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b424669-446e-4b8a-87b4-0bf6d179ec25" containerName="init" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.730304 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8614a49-3244-4947-b65b-bc61a8490e30" containerName="neutron-db-sync" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.730317 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b424669-446e-4b8a-87b4-0bf6d179ec25" containerName="dnsmasq-dns" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.731216 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.796509 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-wbj9t"] Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.810056 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7496d77c9d-kjd5v"] Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.811588 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.814017 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-5788c" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.814621 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.814860 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.817604 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.825111 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7496d77c9d-kjd5v"] Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.880518 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-config\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.880593 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-dns-svc\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.880627 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.880655 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.880712 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.880731 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4qg6\" (UniqueName: \"kubernetes.io/projected/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-kube-api-access-d4qg6\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.981858 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-config\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.981963 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-dns-svc\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.981992 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-combined-ca-bundle\") pod \"neutron-7496d77c9d-kjd5v\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.982025 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs8gf\" (UniqueName: \"kubernetes.io/projected/d6850510-af16-4d30-b30c-54a9bd2186d3-kube-api-access-fs8gf\") pod \"neutron-7496d77c9d-kjd5v\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.982041 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.982064 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-httpd-config\") pod \"neutron-7496d77c9d-kjd5v\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.982094 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.982121 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-config\") pod \"neutron-7496d77c9d-kjd5v\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.982168 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.982193 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4qg6\" (UniqueName: \"kubernetes.io/projected/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-kube-api-access-d4qg6\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.982212 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-ovndb-tls-certs\") pod \"neutron-7496d77c9d-kjd5v\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.983262 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-config\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.983870 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-dns-svc\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.984446 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.985017 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:52 crc kubenswrapper[4806]: I1204 04:14:52.985592 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.018073 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4qg6\" (UniqueName: \"kubernetes.io/projected/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-kube-api-access-d4qg6\") pod \"dnsmasq-dns-55f844cf75-wbj9t\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.083641 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-httpd-config\") pod \"neutron-7496d77c9d-kjd5v\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.083773 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-config\") pod \"neutron-7496d77c9d-kjd5v\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.083868 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-ovndb-tls-certs\") pod \"neutron-7496d77c9d-kjd5v\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.083973 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-combined-ca-bundle\") pod \"neutron-7496d77c9d-kjd5v\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.084016 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs8gf\" (UniqueName: \"kubernetes.io/projected/d6850510-af16-4d30-b30c-54a9bd2186d3-kube-api-access-fs8gf\") pod \"neutron-7496d77c9d-kjd5v\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.091832 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-combined-ca-bundle\") pod \"neutron-7496d77c9d-kjd5v\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.094799 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-httpd-config\") pod \"neutron-7496d77c9d-kjd5v\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.105721 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-config\") pod \"neutron-7496d77c9d-kjd5v\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.106207 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-ovndb-tls-certs\") pod \"neutron-7496d77c9d-kjd5v\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.118466 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.132737 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs8gf\" (UniqueName: \"kubernetes.io/projected/d6850510-af16-4d30-b30c-54a9bd2186d3-kube-api-access-fs8gf\") pod \"neutron-7496d77c9d-kjd5v\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.188214 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.217280 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gpzjd" event={"ID":"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc","Type":"ContainerStarted","Data":"af53fa416c57a054383727c3a586fcbe9eae28466916991d55ae07c02d62462c"} Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.267545 4806 generic.go:334] "Generic (PLEG): container finished" podID="60a725fe-fe27-4d81-826d-d86a501d86ec" containerID="3bc188417bac387dce862e2208d528d50a0a495301490586f28ed4127888b2c9" exitCode=0 Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.267656 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" event={"ID":"60a725fe-fe27-4d81-826d-d86a501d86ec","Type":"ContainerDied","Data":"3bc188417bac387dce862e2208d528d50a0a495301490586f28ed4127888b2c9"} Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.267695 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" event={"ID":"60a725fe-fe27-4d81-826d-d86a501d86ec","Type":"ContainerStarted","Data":"4aeeb9138b00eda926bcfc8fd4bcdbfd8538ea6c526c24603cdb84ea8cb33a6e"} Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.267894 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" podUID="60a725fe-fe27-4d81-826d-d86a501d86ec" containerName="dnsmasq-dns" containerID="cri-o://4aeeb9138b00eda926bcfc8fd4bcdbfd8538ea6c526c24603cdb84ea8cb33a6e" gracePeriod=10 Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.268035 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.338531 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4783b024-15d6-44a8-bf7f-5ae7f52ce45f","Type":"ContainerStarted","Data":"aa18177a72fdef2f8f545ffb843dbf7b28b986c608f90a5543e9f0e9d603633f"} Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.357551 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-gpzjd" podStartSLOduration=9.241404245 podStartE2EDuration="43.35753474s" podCreationTimestamp="2025-12-04 04:14:10 +0000 UTC" firstStartedPulling="2025-12-04 04:14:13.317055415 +0000 UTC m=+1168.175568343" lastFinishedPulling="2025-12-04 04:14:47.43318591 +0000 UTC m=+1202.291698838" observedRunningTime="2025-12-04 04:14:53.304113732 +0000 UTC m=+1208.162626710" watchObservedRunningTime="2025-12-04 04:14:53.35753474 +0000 UTC m=+1208.216047668" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.369258 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" podStartSLOduration=34.369232677 podStartE2EDuration="34.369232677s" podCreationTimestamp="2025-12-04 04:14:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:14:53.35436783 +0000 UTC m=+1208.212880768" watchObservedRunningTime="2025-12-04 04:14:53.369232677 +0000 UTC m=+1208.227745605" Dec 04 04:14:53 crc kubenswrapper[4806]: I1204 04:14:53.379191 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sbp9s" event={"ID":"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36","Type":"ContainerStarted","Data":"19d388c1240f3f615e6c24b37e7ccf33f547b59fa258adddff42e50f7773b7b9"} Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.000798 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-sbp9s" podStartSLOduration=26.000774758 podStartE2EDuration="26.000774758s" podCreationTimestamp="2025-12-04 04:14:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:14:53.440255577 +0000 UTC m=+1208.298768505" watchObservedRunningTime="2025-12-04 04:14:54.000774758 +0000 UTC m=+1208.859287696" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.047699 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7496d77c9d-kjd5v"] Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.076809 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.232499 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-ovsdbserver-sb\") pod \"60a725fe-fe27-4d81-826d-d86a501d86ec\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.232993 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-dns-swift-storage-0\") pod \"60a725fe-fe27-4d81-826d-d86a501d86ec\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.233044 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-config\") pod \"60a725fe-fe27-4d81-826d-d86a501d86ec\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.233136 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g8z6\" (UniqueName: \"kubernetes.io/projected/60a725fe-fe27-4d81-826d-d86a501d86ec-kube-api-access-8g8z6\") pod \"60a725fe-fe27-4d81-826d-d86a501d86ec\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.233187 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-ovsdbserver-nb\") pod \"60a725fe-fe27-4d81-826d-d86a501d86ec\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.233251 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-dns-svc\") pod \"60a725fe-fe27-4d81-826d-d86a501d86ec\" (UID: \"60a725fe-fe27-4d81-826d-d86a501d86ec\") " Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.310775 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60a725fe-fe27-4d81-826d-d86a501d86ec-kube-api-access-8g8z6" (OuterVolumeSpecName: "kube-api-access-8g8z6") pod "60a725fe-fe27-4d81-826d-d86a501d86ec" (UID: "60a725fe-fe27-4d81-826d-d86a501d86ec"). InnerVolumeSpecName "kube-api-access-8g8z6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.317975 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-wbj9t"] Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.335237 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g8z6\" (UniqueName: \"kubernetes.io/projected/60a725fe-fe27-4d81-826d-d86a501d86ec-kube-api-access-8g8z6\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.414428 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6","Type":"ContainerStarted","Data":"0f74dbfa1ec3fa6acfe1c76b9a1e7de981ce6ee33106c9c54d83800de21f0d6f"} Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.452052 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e6c3203e-b396-46d8-b74d-24ed86ff2a60","Type":"ContainerStarted","Data":"c0437542dd5ecba79998bf56d0feaa784401b943bd412a702d983b4aedd1ff31"} Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.475180 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" event={"ID":"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b","Type":"ContainerStarted","Data":"cf6c16c3b16e4c90fc22041149c9f46639f91f7f32f8ff24db640101783a42eb"} Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.506611 4806 generic.go:334] "Generic (PLEG): container finished" podID="60a725fe-fe27-4d81-826d-d86a501d86ec" containerID="4aeeb9138b00eda926bcfc8fd4bcdbfd8538ea6c526c24603cdb84ea8cb33a6e" exitCode=0 Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.506681 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" event={"ID":"60a725fe-fe27-4d81-826d-d86a501d86ec","Type":"ContainerDied","Data":"4aeeb9138b00eda926bcfc8fd4bcdbfd8538ea6c526c24603cdb84ea8cb33a6e"} Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.506713 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" event={"ID":"60a725fe-fe27-4d81-826d-d86a501d86ec","Type":"ContainerDied","Data":"71bd9b6333c48cf446072af8d54a502b65602d999c18b86ecfd75bbaa6507815"} Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.506730 4806 scope.go:117] "RemoveContainer" containerID="4aeeb9138b00eda926bcfc8fd4bcdbfd8538ea6c526c24603cdb84ea8cb33a6e" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.508219 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-s94pm" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.515312 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "60a725fe-fe27-4d81-826d-d86a501d86ec" (UID: "60a725fe-fe27-4d81-826d-d86a501d86ec"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.515636 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7496d77c9d-kjd5v" event={"ID":"d6850510-af16-4d30-b30c-54a9bd2186d3","Type":"ContainerStarted","Data":"7db00b93a79ccf42a6471519ff561df1a65bd2853a28e50441a06a1284641f08"} Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.527154 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "60a725fe-fe27-4d81-826d-d86a501d86ec" (UID: "60a725fe-fe27-4d81-826d-d86a501d86ec"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.530205 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "60a725fe-fe27-4d81-826d-d86a501d86ec" (UID: "60a725fe-fe27-4d81-826d-d86a501d86ec"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.540304 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d8bd6c96d-4sqhf" event={"ID":"d2d53864-c3c8-4e40-8791-0bac49ba561e","Type":"ContainerStarted","Data":"a440aa262b334c36c507fc4d0fece95b286074ca6be74690b931e0ac6569d085"} Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.540351 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d8bd6c96d-4sqhf" event={"ID":"d2d53864-c3c8-4e40-8791-0bac49ba561e","Type":"ContainerStarted","Data":"092967098e2ef8c7603d4afe480336c178a549ddcde98c87b4344631c94fe098"} Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.551919 4806 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.552138 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.552395 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.570700 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "60a725fe-fe27-4d81-826d-d86a501d86ec" (UID: "60a725fe-fe27-4d81-826d-d86a501d86ec"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.577468 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc4f74d8d-xj2xn" event={"ID":"2265394b-be59-4797-b54d-369e5a5624e0","Type":"ContainerStarted","Data":"33669b61f9a2e337d989305386e72fc8b569e4ce6dbdb13e37c0aacc3aaf1ca1"} Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.577508 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc4f74d8d-xj2xn" event={"ID":"2265394b-be59-4797-b54d-369e5a5624e0","Type":"ContainerStarted","Data":"1b52080a09d2ec89dad35352b7fbd08759432c499eaa43fddc058c1d3aa391e5"} Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.582330 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7d8bd6c96d-4sqhf" podStartSLOduration=35.081519303 podStartE2EDuration="35.582312798s" podCreationTimestamp="2025-12-04 04:14:19 +0000 UTC" firstStartedPulling="2025-12-04 04:14:51.81364727 +0000 UTC m=+1206.672160198" lastFinishedPulling="2025-12-04 04:14:52.314440765 +0000 UTC m=+1207.172953693" observedRunningTime="2025-12-04 04:14:54.565336655 +0000 UTC m=+1209.423849573" watchObservedRunningTime="2025-12-04 04:14:54.582312798 +0000 UTC m=+1209.440825726" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.615703 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-config" (OuterVolumeSpecName: "config") pod "60a725fe-fe27-4d81-826d-d86a501d86ec" (UID: "60a725fe-fe27-4d81-826d-d86a501d86ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.617668 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7bc4f74d8d-xj2xn" podStartSLOduration=35.115904292 podStartE2EDuration="35.617645688s" podCreationTimestamp="2025-12-04 04:14:19 +0000 UTC" firstStartedPulling="2025-12-04 04:14:51.813728382 +0000 UTC m=+1206.672241310" lastFinishedPulling="2025-12-04 04:14:52.315469778 +0000 UTC m=+1207.173982706" observedRunningTime="2025-12-04 04:14:54.603708421 +0000 UTC m=+1209.462221349" watchObservedRunningTime="2025-12-04 04:14:54.617645688 +0000 UTC m=+1209.476158616" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.653651 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.653684 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60a725fe-fe27-4d81-826d-d86a501d86ec-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.676089 4806 scope.go:117] "RemoveContainer" containerID="3bc188417bac387dce862e2208d528d50a0a495301490586f28ed4127888b2c9" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.761564 4806 scope.go:117] "RemoveContainer" containerID="4aeeb9138b00eda926bcfc8fd4bcdbfd8538ea6c526c24603cdb84ea8cb33a6e" Dec 04 04:14:54 crc kubenswrapper[4806]: E1204 04:14:54.762494 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aeeb9138b00eda926bcfc8fd4bcdbfd8538ea6c526c24603cdb84ea8cb33a6e\": container with ID starting with 4aeeb9138b00eda926bcfc8fd4bcdbfd8538ea6c526c24603cdb84ea8cb33a6e not found: ID does not exist" containerID="4aeeb9138b00eda926bcfc8fd4bcdbfd8538ea6c526c24603cdb84ea8cb33a6e" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.762536 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aeeb9138b00eda926bcfc8fd4bcdbfd8538ea6c526c24603cdb84ea8cb33a6e"} err="failed to get container status \"4aeeb9138b00eda926bcfc8fd4bcdbfd8538ea6c526c24603cdb84ea8cb33a6e\": rpc error: code = NotFound desc = could not find container \"4aeeb9138b00eda926bcfc8fd4bcdbfd8538ea6c526c24603cdb84ea8cb33a6e\": container with ID starting with 4aeeb9138b00eda926bcfc8fd4bcdbfd8538ea6c526c24603cdb84ea8cb33a6e not found: ID does not exist" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.762562 4806 scope.go:117] "RemoveContainer" containerID="3bc188417bac387dce862e2208d528d50a0a495301490586f28ed4127888b2c9" Dec 04 04:14:54 crc kubenswrapper[4806]: E1204 04:14:54.762832 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bc188417bac387dce862e2208d528d50a0a495301490586f28ed4127888b2c9\": container with ID starting with 3bc188417bac387dce862e2208d528d50a0a495301490586f28ed4127888b2c9 not found: ID does not exist" containerID="3bc188417bac387dce862e2208d528d50a0a495301490586f28ed4127888b2c9" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.762862 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bc188417bac387dce862e2208d528d50a0a495301490586f28ed4127888b2c9"} err="failed to get container status \"3bc188417bac387dce862e2208d528d50a0a495301490586f28ed4127888b2c9\": rpc error: code = NotFound desc = could not find container \"3bc188417bac387dce862e2208d528d50a0a495301490586f28ed4127888b2c9\": container with ID starting with 3bc188417bac387dce862e2208d528d50a0a495301490586f28ed4127888b2c9 not found: ID does not exist" Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.867223 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-s94pm"] Dec 04 04:14:54 crc kubenswrapper[4806]: I1204 04:14:54.883831 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-s94pm"] Dec 04 04:14:55 crc kubenswrapper[4806]: I1204 04:14:55.439725 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60a725fe-fe27-4d81-826d-d86a501d86ec" path="/var/lib/kubelet/pods/60a725fe-fe27-4d81-826d-d86a501d86ec/volumes" Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.615264 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6","Type":"ContainerStarted","Data":"30e37490ab0ae0f6c1f60e0e0e621a113bb6515ea709a9445c8f51d24b08ce98"} Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.615616 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" containerName="glance-log" containerID="cri-o://0f74dbfa1ec3fa6acfe1c76b9a1e7de981ce6ee33106c9c54d83800de21f0d6f" gracePeriod=30 Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.615730 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" containerName="glance-httpd" containerID="cri-o://30e37490ab0ae0f6c1f60e0e0e621a113bb6515ea709a9445c8f51d24b08ce98" gracePeriod=30 Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.629873 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e6c3203e-b396-46d8-b74d-24ed86ff2a60" containerName="glance-log" containerID="cri-o://c0437542dd5ecba79998bf56d0feaa784401b943bd412a702d983b4aedd1ff31" gracePeriod=30 Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.630003 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e6c3203e-b396-46d8-b74d-24ed86ff2a60","Type":"ContainerStarted","Data":"49f46582b8b573c0497d073a77894ac0a088bb4fc485a82d7aa2519f9c80a95e"} Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.630023 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e6c3203e-b396-46d8-b74d-24ed86ff2a60" containerName="glance-httpd" containerID="cri-o://49f46582b8b573c0497d073a77894ac0a088bb4fc485a82d7aa2519f9c80a95e" gracePeriod=30 Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.642374 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=37.642352755 podStartE2EDuration="37.642352755s" podCreationTimestamp="2025-12-04 04:14:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:14:56.641885381 +0000 UTC m=+1211.500398309" watchObservedRunningTime="2025-12-04 04:14:56.642352755 +0000 UTC m=+1211.500865693" Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.643718 4806 generic.go:334] "Generic (PLEG): container finished" podID="88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b" containerID="f9cd5f81df81499f791ff879391331a78bfc1dbd5a2e9964787d94f66ec4edfc" exitCode=0 Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.643795 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" event={"ID":"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b","Type":"ContainerDied","Data":"f9cd5f81df81499f791ff879391331a78bfc1dbd5a2e9964787d94f66ec4edfc"} Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.654571 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7496d77c9d-kjd5v" event={"ID":"d6850510-af16-4d30-b30c-54a9bd2186d3","Type":"ContainerStarted","Data":"96112656a08459e761509553df8027474648735003ea0089e07404ec72b6f05c"} Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.676216 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=37.676194088 podStartE2EDuration="37.676194088s" podCreationTimestamp="2025-12-04 04:14:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:14:56.666911917 +0000 UTC m=+1211.525424855" watchObservedRunningTime="2025-12-04 04:14:56.676194088 +0000 UTC m=+1211.534707016" Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.941339 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-669cc9c8df-p88wk"] Dec 04 04:14:56 crc kubenswrapper[4806]: E1204 04:14:56.941788 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60a725fe-fe27-4d81-826d-d86a501d86ec" containerName="init" Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.941810 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="60a725fe-fe27-4d81-826d-d86a501d86ec" containerName="init" Dec 04 04:14:56 crc kubenswrapper[4806]: E1204 04:14:56.941852 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60a725fe-fe27-4d81-826d-d86a501d86ec" containerName="dnsmasq-dns" Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.941859 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="60a725fe-fe27-4d81-826d-d86a501d86ec" containerName="dnsmasq-dns" Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.942128 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="60a725fe-fe27-4d81-826d-d86a501d86ec" containerName="dnsmasq-dns" Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.943348 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.946273 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.946594 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 04 04:14:56 crc kubenswrapper[4806]: I1204 04:14:56.973794 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-669cc9c8df-p88wk"] Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.015804 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-combined-ca-bundle\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.015887 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-internal-tls-certs\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.015949 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-httpd-config\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.016190 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-ovndb-tls-certs\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.016230 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-config\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.016255 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6mpp\" (UniqueName: \"kubernetes.io/projected/7fbb687c-667a-4025-bd3c-1eb037758c0d-kube-api-access-s6mpp\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.016272 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-public-tls-certs\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.124127 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-combined-ca-bundle\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.124237 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-internal-tls-certs\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.124293 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-httpd-config\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.124326 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-ovndb-tls-certs\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.124376 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-config\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.124427 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6mpp\" (UniqueName: \"kubernetes.io/projected/7fbb687c-667a-4025-bd3c-1eb037758c0d-kube-api-access-s6mpp\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.124448 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-public-tls-certs\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.133233 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-httpd-config\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.137817 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-internal-tls-certs\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.145825 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-public-tls-certs\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.146526 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-combined-ca-bundle\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.147298 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-config\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.142308 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fbb687c-667a-4025-bd3c-1eb037758c0d-ovndb-tls-certs\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.152888 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6mpp\" (UniqueName: \"kubernetes.io/projected/7fbb687c-667a-4025-bd3c-1eb037758c0d-kube-api-access-s6mpp\") pod \"neutron-669cc9c8df-p88wk\" (UID: \"7fbb687c-667a-4025-bd3c-1eb037758c0d\") " pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.267766 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.682589 4806 generic.go:334] "Generic (PLEG): container finished" podID="e6c3203e-b396-46d8-b74d-24ed86ff2a60" containerID="49f46582b8b573c0497d073a77894ac0a088bb4fc485a82d7aa2519f9c80a95e" exitCode=143 Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.682621 4806 generic.go:334] "Generic (PLEG): container finished" podID="e6c3203e-b396-46d8-b74d-24ed86ff2a60" containerID="c0437542dd5ecba79998bf56d0feaa784401b943bd412a702d983b4aedd1ff31" exitCode=143 Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.682683 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e6c3203e-b396-46d8-b74d-24ed86ff2a60","Type":"ContainerDied","Data":"49f46582b8b573c0497d073a77894ac0a088bb4fc485a82d7aa2519f9c80a95e"} Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.682753 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e6c3203e-b396-46d8-b74d-24ed86ff2a60","Type":"ContainerDied","Data":"c0437542dd5ecba79998bf56d0feaa784401b943bd412a702d983b4aedd1ff31"} Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.693265 4806 generic.go:334] "Generic (PLEG): container finished" podID="0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" containerID="30e37490ab0ae0f6c1f60e0e0e621a113bb6515ea709a9445c8f51d24b08ce98" exitCode=143 Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.693306 4806 generic.go:334] "Generic (PLEG): container finished" podID="0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" containerID="0f74dbfa1ec3fa6acfe1c76b9a1e7de981ce6ee33106c9c54d83800de21f0d6f" exitCode=143 Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.693332 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6","Type":"ContainerDied","Data":"30e37490ab0ae0f6c1f60e0e0e621a113bb6515ea709a9445c8f51d24b08ce98"} Dec 04 04:14:57 crc kubenswrapper[4806]: I1204 04:14:57.693364 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6","Type":"ContainerDied","Data":"0f74dbfa1ec3fa6acfe1c76b9a1e7de981ce6ee33106c9c54d83800de21f0d6f"} Dec 04 04:14:59 crc kubenswrapper[4806]: I1204 04:14:59.710135 4806 generic.go:334] "Generic (PLEG): container finished" podID="a8047cee-ef2b-46ab-aca7-faebe1ae8dfc" containerID="af53fa416c57a054383727c3a586fcbe9eae28466916991d55ae07c02d62462c" exitCode=0 Dec 04 04:14:59 crc kubenswrapper[4806]: I1204 04:14:59.710279 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gpzjd" event={"ID":"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc","Type":"ContainerDied","Data":"af53fa416c57a054383727c3a586fcbe9eae28466916991d55ae07c02d62462c"} Dec 04 04:14:59 crc kubenswrapper[4806]: I1204 04:14:59.943073 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:14:59 crc kubenswrapper[4806]: I1204 04:14:59.943123 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.156955 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq"] Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.160734 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.163038 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.163085 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.179312 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq"] Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.280772 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d855714a-f641-44bc-9b95-ebda879968b0-config-volume\") pod \"collect-profiles-29413695-thgjq\" (UID: \"d855714a-f641-44bc-9b95-ebda879968b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.280874 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d855714a-f641-44bc-9b95-ebda879968b0-secret-volume\") pod \"collect-profiles-29413695-thgjq\" (UID: \"d855714a-f641-44bc-9b95-ebda879968b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.281441 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss7r7\" (UniqueName: \"kubernetes.io/projected/d855714a-f641-44bc-9b95-ebda879968b0-kube-api-access-ss7r7\") pod \"collect-profiles-29413695-thgjq\" (UID: \"d855714a-f641-44bc-9b95-ebda879968b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.376211 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.376264 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.382856 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d855714a-f641-44bc-9b95-ebda879968b0-config-volume\") pod \"collect-profiles-29413695-thgjq\" (UID: \"d855714a-f641-44bc-9b95-ebda879968b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.382924 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d855714a-f641-44bc-9b95-ebda879968b0-secret-volume\") pod \"collect-profiles-29413695-thgjq\" (UID: \"d855714a-f641-44bc-9b95-ebda879968b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.383048 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss7r7\" (UniqueName: \"kubernetes.io/projected/d855714a-f641-44bc-9b95-ebda879968b0-kube-api-access-ss7r7\") pod \"collect-profiles-29413695-thgjq\" (UID: \"d855714a-f641-44bc-9b95-ebda879968b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.383818 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d855714a-f641-44bc-9b95-ebda879968b0-config-volume\") pod \"collect-profiles-29413695-thgjq\" (UID: \"d855714a-f641-44bc-9b95-ebda879968b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.397308 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d855714a-f641-44bc-9b95-ebda879968b0-secret-volume\") pod \"collect-profiles-29413695-thgjq\" (UID: \"d855714a-f641-44bc-9b95-ebda879968b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.398965 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss7r7\" (UniqueName: \"kubernetes.io/projected/d855714a-f641-44bc-9b95-ebda879968b0-kube-api-access-ss7r7\") pod \"collect-profiles-29413695-thgjq\" (UID: \"d855714a-f641-44bc-9b95-ebda879968b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" Dec 04 04:15:00 crc kubenswrapper[4806]: I1204 04:15:00.492199 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.547674 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gpzjd" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.593218 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.638017 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-httpd-run\") pod \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.638119 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psgkc\" (UniqueName: \"kubernetes.io/projected/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-kube-api-access-psgkc\") pod \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.638148 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-config-data\") pod \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.638187 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-scripts\") pod \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.638236 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-config-data\") pod \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.638275 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99g7p\" (UniqueName: \"kubernetes.io/projected/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-kube-api-access-99g7p\") pod \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.638304 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-logs\") pod \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.638345 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-combined-ca-bundle\") pod \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.638387 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-scripts\") pod \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\" (UID: \"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc\") " Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.638414 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.638437 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-logs\") pod \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.638549 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-combined-ca-bundle\") pod \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\" (UID: \"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6\") " Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.640746 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" (UID: "0ba8ca2b-670f-4f0d-927c-ff8959ef65d6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.641045 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-logs" (OuterVolumeSpecName: "logs") pod "a8047cee-ef2b-46ab-aca7-faebe1ae8dfc" (UID: "a8047cee-ef2b-46ab-aca7-faebe1ae8dfc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.646982 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-logs" (OuterVolumeSpecName: "logs") pod "0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" (UID: "0ba8ca2b-670f-4f0d-927c-ff8959ef65d6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.650820 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-scripts" (OuterVolumeSpecName: "scripts") pod "0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" (UID: "0ba8ca2b-670f-4f0d-927c-ff8959ef65d6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.659444 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-kube-api-access-99g7p" (OuterVolumeSpecName: "kube-api-access-99g7p") pod "a8047cee-ef2b-46ab-aca7-faebe1ae8dfc" (UID: "a8047cee-ef2b-46ab-aca7-faebe1ae8dfc"). InnerVolumeSpecName "kube-api-access-99g7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.659636 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-scripts" (OuterVolumeSpecName: "scripts") pod "a8047cee-ef2b-46ab-aca7-faebe1ae8dfc" (UID: "a8047cee-ef2b-46ab-aca7-faebe1ae8dfc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.676910 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" (UID: "0ba8ca2b-670f-4f0d-927c-ff8959ef65d6"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.676954 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-kube-api-access-psgkc" (OuterVolumeSpecName: "kube-api-access-psgkc") pod "0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" (UID: "0ba8ca2b-670f-4f0d-927c-ff8959ef65d6"). InnerVolumeSpecName "kube-api-access-psgkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.741377 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psgkc\" (UniqueName: \"kubernetes.io/projected/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-kube-api-access-psgkc\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.741414 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.741427 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99g7p\" (UniqueName: \"kubernetes.io/projected/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-kube-api-access-99g7p\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.741437 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.741447 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.741495 4806 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.741507 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.741518 4806 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.756487 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gpzjd" event={"ID":"a8047cee-ef2b-46ab-aca7-faebe1ae8dfc","Type":"ContainerDied","Data":"765aa2df37d33eea27b8135b297249c3b1f31f37ef6af6b7731df8de975ca976"} Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.756534 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="765aa2df37d33eea27b8135b297249c3b1f31f37ef6af6b7731df8de975ca976" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.756710 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gpzjd" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.758125 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" (UID: "0ba8ca2b-670f-4f0d-927c-ff8959ef65d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.762431 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0ba8ca2b-670f-4f0d-927c-ff8959ef65d6","Type":"ContainerDied","Data":"f783b666c8119ac094f3c24bca0cb7252cb1ac304bc0746160703daa8a581f45"} Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.762475 4806 scope.go:117] "RemoveContainer" containerID="30e37490ab0ae0f6c1f60e0e0e621a113bb6515ea709a9445c8f51d24b08ce98" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.762606 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.839172 4806 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.843401 4806 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.843440 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.860228 4806 scope.go:117] "RemoveContainer" containerID="0f74dbfa1ec3fa6acfe1c76b9a1e7de981ce6ee33106c9c54d83800de21f0d6f" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.974900 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8047cee-ef2b-46ab-aca7-faebe1ae8dfc" (UID: "a8047cee-ef2b-46ab-aca7-faebe1ae8dfc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:02 crc kubenswrapper[4806]: I1204 04:15:02.980533 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-config-data" (OuterVolumeSpecName: "config-data") pod "a8047cee-ef2b-46ab-aca7-faebe1ae8dfc" (UID: "a8047cee-ef2b-46ab-aca7-faebe1ae8dfc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:02.998577 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-config-data" (OuterVolumeSpecName: "config-data") pod "0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" (UID: "0ba8ca2b-670f-4f0d-927c-ff8959ef65d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.048524 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.048563 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.048573 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.245602 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq"] Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.266277 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.365440 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6c3203e-b396-46d8-b74d-24ed86ff2a60-httpd-run\") pod \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.365946 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6c3203e-b396-46d8-b74d-24ed86ff2a60-logs\") pod \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.365982 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqk9j\" (UniqueName: \"kubernetes.io/projected/e6c3203e-b396-46d8-b74d-24ed86ff2a60-kube-api-access-mqk9j\") pod \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.366076 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.366136 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-config-data\") pod \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.366196 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-combined-ca-bundle\") pod \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.366235 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-scripts\") pod \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\" (UID: \"e6c3203e-b396-46d8-b74d-24ed86ff2a60\") " Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.368319 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6c3203e-b396-46d8-b74d-24ed86ff2a60-logs" (OuterVolumeSpecName: "logs") pod "e6c3203e-b396-46d8-b74d-24ed86ff2a60" (UID: "e6c3203e-b396-46d8-b74d-24ed86ff2a60"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.368433 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6c3203e-b396-46d8-b74d-24ed86ff2a60-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e6c3203e-b396-46d8-b74d-24ed86ff2a60" (UID: "e6c3203e-b396-46d8-b74d-24ed86ff2a60"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.368889 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.392327 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.395607 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6c3203e-b396-46d8-b74d-24ed86ff2a60-kube-api-access-mqk9j" (OuterVolumeSpecName: "kube-api-access-mqk9j") pod "e6c3203e-b396-46d8-b74d-24ed86ff2a60" (UID: "e6c3203e-b396-46d8-b74d-24ed86ff2a60"). InnerVolumeSpecName "kube-api-access-mqk9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.395761 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-scripts" (OuterVolumeSpecName: "scripts") pod "e6c3203e-b396-46d8-b74d-24ed86ff2a60" (UID: "e6c3203e-b396-46d8-b74d-24ed86ff2a60"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.395874 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "e6c3203e-b396-46d8-b74d-24ed86ff2a60" (UID: "e6c3203e-b396-46d8-b74d-24ed86ff2a60"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.454314 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" path="/var/lib/kubelet/pods/0ba8ca2b-670f-4f0d-927c-ff8959ef65d6/volumes" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.464754 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-669cc9c8df-p88wk"] Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.468884 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6c3203e-b396-46d8-b74d-24ed86ff2a60-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.468915 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqk9j\" (UniqueName: \"kubernetes.io/projected/e6c3203e-b396-46d8-b74d-24ed86ff2a60-kube-api-access-mqk9j\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.469073 4806 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.469113 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.469124 4806 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e6c3203e-b396-46d8-b74d-24ed86ff2a60-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.477069 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 04:15:03 crc kubenswrapper[4806]: E1204 04:15:03.477503 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6c3203e-b396-46d8-b74d-24ed86ff2a60" containerName="glance-log" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.477522 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6c3203e-b396-46d8-b74d-24ed86ff2a60" containerName="glance-log" Dec 04 04:15:03 crc kubenswrapper[4806]: E1204 04:15:03.477533 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" containerName="glance-log" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.477542 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" containerName="glance-log" Dec 04 04:15:03 crc kubenswrapper[4806]: E1204 04:15:03.477559 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" containerName="glance-httpd" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.477565 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" containerName="glance-httpd" Dec 04 04:15:03 crc kubenswrapper[4806]: E1204 04:15:03.477578 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6c3203e-b396-46d8-b74d-24ed86ff2a60" containerName="glance-httpd" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.477584 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6c3203e-b396-46d8-b74d-24ed86ff2a60" containerName="glance-httpd" Dec 04 04:15:03 crc kubenswrapper[4806]: E1204 04:15:03.477592 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8047cee-ef2b-46ab-aca7-faebe1ae8dfc" containerName="placement-db-sync" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.477599 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8047cee-ef2b-46ab-aca7-faebe1ae8dfc" containerName="placement-db-sync" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.477770 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8047cee-ef2b-46ab-aca7-faebe1ae8dfc" containerName="placement-db-sync" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.477788 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6c3203e-b396-46d8-b74d-24ed86ff2a60" containerName="glance-httpd" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.477798 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6c3203e-b396-46d8-b74d-24ed86ff2a60" containerName="glance-log" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.477813 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" containerName="glance-log" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.477824 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ba8ca2b-670f-4f0d-927c-ff8959ef65d6" containerName="glance-httpd" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.478768 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.482034 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e6c3203e-b396-46d8-b74d-24ed86ff2a60" (UID: "e6c3203e-b396-46d8-b74d-24ed86ff2a60"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.482440 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.497968 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.498221 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.567447 4806 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.570324 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/60587700-0430-4cc5-a572-234a2a946748-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.570439 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60587700-0430-4cc5-a572-234a2a946748-logs\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.570547 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-scripts\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.570673 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.570778 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-config-data\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.570913 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.571038 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxkjq\" (UniqueName: \"kubernetes.io/projected/60587700-0430-4cc5-a572-234a2a946748-kube-api-access-qxkjq\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.571148 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.571261 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.571317 4806 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.609234 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-config-data" (OuterVolumeSpecName: "config-data") pod "e6c3203e-b396-46d8-b74d-24ed86ff2a60" (UID: "e6c3203e-b396-46d8-b74d-24ed86ff2a60"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.671378 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-d9c876cd8-zfctt"] Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.672455 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxkjq\" (UniqueName: \"kubernetes.io/projected/60587700-0430-4cc5-a572-234a2a946748-kube-api-access-qxkjq\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.672577 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.672685 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/60587700-0430-4cc5-a572-234a2a946748-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.672769 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60587700-0430-4cc5-a572-234a2a946748-logs\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.672842 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-scripts\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.672916 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.673043 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-config-data\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.673138 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.673244 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6c3203e-b396-46d8-b74d-24ed86ff2a60-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.673174 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.674192 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.674626 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/60587700-0430-4cc5-a572-234a2a946748-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.674962 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60587700-0430-4cc5-a572-234a2a946748-logs\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.683594 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-scripts\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.683738 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.686945 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.687265 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.691300 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.691604 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vvlq5" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.691665 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.692775 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-config-data\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.702118 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.705075 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d9c876cd8-zfctt"] Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.712983 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxkjq\" (UniqueName: \"kubernetes.io/projected/60587700-0430-4cc5-a572-234a2a946748-kube-api-access-qxkjq\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.761646 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.776455 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122ae99a-2891-4bdb-903a-9cccfb3df2a1-config-data\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.776514 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122ae99a-2891-4bdb-903a-9cccfb3df2a1-logs\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.776557 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/122ae99a-2891-4bdb-903a-9cccfb3df2a1-internal-tls-certs\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.776592 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdr9r\" (UniqueName: \"kubernetes.io/projected/122ae99a-2891-4bdb-903a-9cccfb3df2a1-kube-api-access-jdr9r\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.776618 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/122ae99a-2891-4bdb-903a-9cccfb3df2a1-public-tls-certs\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.776696 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122ae99a-2891-4bdb-903a-9cccfb3df2a1-combined-ca-bundle\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.776721 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/122ae99a-2891-4bdb-903a-9cccfb3df2a1-scripts\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.786780 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pl5fx" event={"ID":"b1e9cfb9-1741-4a4e-a05d-7d666699876c","Type":"ContainerStarted","Data":"28cf550b5b6b5bf32045b1666c1c389a131dee767915bf17171311afddd538f3"} Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.794597 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7496d77c9d-kjd5v" event={"ID":"d6850510-af16-4d30-b30c-54a9bd2186d3","Type":"ContainerStarted","Data":"95dd0182bcbefacae5d7c5b8075c03f37fcb2e14fb34b4ee557b3091755a4825"} Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.794788 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.802356 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.802578 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4783b024-15d6-44a8-bf7f-5ae7f52ce45f","Type":"ContainerStarted","Data":"c12f9444c959f7c0ad73200475b9e50f02ccf11ba247694c417711e06c86f43e"} Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.821544 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" event={"ID":"d855714a-f641-44bc-9b95-ebda879968b0","Type":"ContainerStarted","Data":"581a1e4deb008f3479806b535a070be534b5d91c291b02caf9d8ede622d3103b"} Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.821584 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" event={"ID":"d855714a-f641-44bc-9b95-ebda879968b0","Type":"ContainerStarted","Data":"d20bc5f45571f0bad04fcea30592e462fbddc1ed891a0f6082ea85250ab4a02a"} Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.834359 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-pl5fx" podStartSLOduration=3.466299361 podStartE2EDuration="53.834342582s" podCreationTimestamp="2025-12-04 04:14:10 +0000 UTC" firstStartedPulling="2025-12-04 04:14:12.649944196 +0000 UTC m=+1167.508457124" lastFinishedPulling="2025-12-04 04:15:03.017987417 +0000 UTC m=+1217.876500345" observedRunningTime="2025-12-04 04:15:03.809285295 +0000 UTC m=+1218.667798223" watchObservedRunningTime="2025-12-04 04:15:03.834342582 +0000 UTC m=+1218.692855510" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.840770 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e6c3203e-b396-46d8-b74d-24ed86ff2a60","Type":"ContainerDied","Data":"b5d23decab701a5746d790c14f8b36bc1c010e37ba85829fe55907304998e58b"} Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.840825 4806 scope.go:117] "RemoveContainer" containerID="49f46582b8b573c0497d073a77894ac0a088bb4fc485a82d7aa2519f9c80a95e" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.840994 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.844755 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7496d77c9d-kjd5v" podStartSLOduration=11.844738409 podStartE2EDuration="11.844738409s" podCreationTimestamp="2025-12-04 04:14:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:15:03.833688042 +0000 UTC m=+1218.692200970" watchObservedRunningTime="2025-12-04 04:15:03.844738409 +0000 UTC m=+1218.703251337" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.867525 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" event={"ID":"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b","Type":"ContainerStarted","Data":"f5963e6d29ec3803ccbf9650a1ebe9d814b2db349acc32fccccdc1b1278f6bbe"} Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.868542 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.872204 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" podStartSLOduration=3.87218508 podStartE2EDuration="3.87218508s" podCreationTimestamp="2025-12-04 04:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:15:03.859071199 +0000 UTC m=+1218.717584127" watchObservedRunningTime="2025-12-04 04:15:03.87218508 +0000 UTC m=+1218.730698008" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.878054 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122ae99a-2891-4bdb-903a-9cccfb3df2a1-combined-ca-bundle\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.878143 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/122ae99a-2891-4bdb-903a-9cccfb3df2a1-scripts\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.878189 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122ae99a-2891-4bdb-903a-9cccfb3df2a1-config-data\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.878214 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122ae99a-2891-4bdb-903a-9cccfb3df2a1-logs\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.878282 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/122ae99a-2891-4bdb-903a-9cccfb3df2a1-internal-tls-certs\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.878316 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdr9r\" (UniqueName: \"kubernetes.io/projected/122ae99a-2891-4bdb-903a-9cccfb3df2a1-kube-api-access-jdr9r\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.878338 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/122ae99a-2891-4bdb-903a-9cccfb3df2a1-public-tls-certs\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.884972 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/122ae99a-2891-4bdb-903a-9cccfb3df2a1-logs\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.888469 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-669cc9c8df-p88wk" event={"ID":"7fbb687c-667a-4025-bd3c-1eb037758c0d","Type":"ContainerStarted","Data":"f0eb4e9910744a3ae2d00fe145520999e9f89a6b06fe359c80e7ad34c419b84e"} Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.900124 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/122ae99a-2891-4bdb-903a-9cccfb3df2a1-scripts\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.901125 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122ae99a-2891-4bdb-903a-9cccfb3df2a1-config-data\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.901370 4806 scope.go:117] "RemoveContainer" containerID="c0437542dd5ecba79998bf56d0feaa784401b943bd412a702d983b4aedd1ff31" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.902335 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122ae99a-2891-4bdb-903a-9cccfb3df2a1-combined-ca-bundle\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.903090 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/122ae99a-2891-4bdb-903a-9cccfb3df2a1-public-tls-certs\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.916677 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/122ae99a-2891-4bdb-903a-9cccfb3df2a1-internal-tls-certs\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.921783 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" podStartSLOduration=11.921761387 podStartE2EDuration="11.921761387s" podCreationTimestamp="2025-12-04 04:14:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:15:03.888780861 +0000 UTC m=+1218.747293789" watchObservedRunningTime="2025-12-04 04:15:03.921761387 +0000 UTC m=+1218.780274315" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.922834 4806 generic.go:334] "Generic (PLEG): container finished" podID="babd629a-b0d5-4ba9-9e01-d36f6bc9ac36" containerID="19d388c1240f3f615e6c24b37e7ccf33f547b59fa258adddff42e50f7773b7b9" exitCode=0 Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.922961 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sbp9s" event={"ID":"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36","Type":"ContainerDied","Data":"19d388c1240f3f615e6c24b37e7ccf33f547b59fa258adddff42e50f7773b7b9"} Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.923773 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdr9r\" (UniqueName: \"kubernetes.io/projected/122ae99a-2891-4bdb-903a-9cccfb3df2a1-kube-api-access-jdr9r\") pod \"placement-d9c876cd8-zfctt\" (UID: \"122ae99a-2891-4bdb-903a-9cccfb3df2a1\") " pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.940727 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.969896 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.985988 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.987480 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.995275 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 04 04:15:03 crc kubenswrapper[4806]: I1204 04:15:03.995538 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.025651 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.073318 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.104339 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.104481 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-logs\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.104563 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.104694 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.104830 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.105055 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.105126 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.105251 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfwnk\" (UniqueName: \"kubernetes.io/projected/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-kube-api-access-lfwnk\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.211131 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.211224 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.211245 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.211289 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfwnk\" (UniqueName: \"kubernetes.io/projected/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-kube-api-access-lfwnk\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.211314 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.211337 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-logs\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.211353 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.211388 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.211874 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.248975 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-logs\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.249214 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.256671 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.259183 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.268098 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.275805 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.298812 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfwnk\" (UniqueName: \"kubernetes.io/projected/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-kube-api-access-lfwnk\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.419474 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.531903 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 04:15:04 crc kubenswrapper[4806]: I1204 04:15:04.841228 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 04:15:04 crc kubenswrapper[4806]: W1204 04:15:04.881915 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60587700_0430_4cc5_a572_234a2a946748.slice/crio-50ea35ebb62b4e8d15ed01641d880bd6f0828594ae397cee91fef8794f4e4425 WatchSource:0}: Error finding container 50ea35ebb62b4e8d15ed01641d880bd6f0828594ae397cee91fef8794f4e4425: Status 404 returned error can't find the container with id 50ea35ebb62b4e8d15ed01641d880bd6f0828594ae397cee91fef8794f4e4425 Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.023260 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-669cc9c8df-p88wk" event={"ID":"7fbb687c-667a-4025-bd3c-1eb037758c0d","Type":"ContainerStarted","Data":"58d8d96b67ad91dd52ec5f9045f70a49a31d652e98d8dfcfb5197d8dcdce2595"} Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.023306 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-669cc9c8df-p88wk" event={"ID":"7fbb687c-667a-4025-bd3c-1eb037758c0d","Type":"ContainerStarted","Data":"7b5e61d21362f7839be9a32fc8547ba4f84939271e2603cd6cba0011c8ae9bd7"} Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.024373 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.033633 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"60587700-0430-4cc5-a572-234a2a946748","Type":"ContainerStarted","Data":"50ea35ebb62b4e8d15ed01641d880bd6f0828594ae397cee91fef8794f4e4425"} Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.038161 4806 generic.go:334] "Generic (PLEG): container finished" podID="d855714a-f641-44bc-9b95-ebda879968b0" containerID="581a1e4deb008f3479806b535a070be534b5d91c291b02caf9d8ede622d3103b" exitCode=0 Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.038236 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" event={"ID":"d855714a-f641-44bc-9b95-ebda879968b0","Type":"ContainerDied","Data":"581a1e4deb008f3479806b535a070be534b5d91c291b02caf9d8ede622d3103b"} Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.054655 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-669cc9c8df-p88wk" podStartSLOduration=9.05462062 podStartE2EDuration="9.05462062s" podCreationTimestamp="2025-12-04 04:14:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:15:05.050750278 +0000 UTC m=+1219.909263206" watchObservedRunningTime="2025-12-04 04:15:05.05462062 +0000 UTC m=+1219.913133548" Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.205648 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d9c876cd8-zfctt"] Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.476977 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6c3203e-b396-46d8-b74d-24ed86ff2a60" path="/var/lib/kubelet/pods/e6c3203e-b396-46d8-b74d-24ed86ff2a60/volumes" Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.626316 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.637092 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:15:05 crc kubenswrapper[4806]: W1204 04:15:05.668677 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06eb71d9_0ccb_4cee_8217_97dae82cbcdf.slice/crio-8bdbdd7a083cb0978851ed58cc23dffb52dd14e942e37541a37efa80260ebbff WatchSource:0}: Error finding container 8bdbdd7a083cb0978851ed58cc23dffb52dd14e942e37541a37efa80260ebbff: Status 404 returned error can't find the container with id 8bdbdd7a083cb0978851ed58cc23dffb52dd14e942e37541a37efa80260ebbff Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.673966 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-config-data\") pod \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.674008 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96s2n\" (UniqueName: \"kubernetes.io/projected/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-kube-api-access-96s2n\") pod \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.674040 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-scripts\") pod \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.674210 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-fernet-keys\") pod \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.674231 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-credential-keys\") pod \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.674295 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-combined-ca-bundle\") pod \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\" (UID: \"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36\") " Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.697170 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "babd629a-b0d5-4ba9-9e01-d36f6bc9ac36" (UID: "babd629a-b0d5-4ba9-9e01-d36f6bc9ac36"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.715431 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-scripts" (OuterVolumeSpecName: "scripts") pod "babd629a-b0d5-4ba9-9e01-d36f6bc9ac36" (UID: "babd629a-b0d5-4ba9-9e01-d36f6bc9ac36"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.715573 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "babd629a-b0d5-4ba9-9e01-d36f6bc9ac36" (UID: "babd629a-b0d5-4ba9-9e01-d36f6bc9ac36"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.716195 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-kube-api-access-96s2n" (OuterVolumeSpecName: "kube-api-access-96s2n") pod "babd629a-b0d5-4ba9-9e01-d36f6bc9ac36" (UID: "babd629a-b0d5-4ba9-9e01-d36f6bc9ac36"). InnerVolumeSpecName "kube-api-access-96s2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.737706 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "babd629a-b0d5-4ba9-9e01-d36f6bc9ac36" (UID: "babd629a-b0d5-4ba9-9e01-d36f6bc9ac36"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.740525 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-config-data" (OuterVolumeSpecName: "config-data") pod "babd629a-b0d5-4ba9-9e01-d36f6bc9ac36" (UID: "babd629a-b0d5-4ba9-9e01-d36f6bc9ac36"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.776795 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.776837 4806 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.776846 4806 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.776855 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.776863 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:05 crc kubenswrapper[4806]: I1204 04:15:05.776872 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96s2n\" (UniqueName: \"kubernetes.io/projected/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36-kube-api-access-96s2n\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.121040 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-64cc7bf6f4-fcxhm"] Dec 04 04:15:06 crc kubenswrapper[4806]: E1204 04:15:06.123269 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="babd629a-b0d5-4ba9-9e01-d36f6bc9ac36" containerName="keystone-bootstrap" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.123294 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="babd629a-b0d5-4ba9-9e01-d36f6bc9ac36" containerName="keystone-bootstrap" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.124460 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="babd629a-b0d5-4ba9-9e01-d36f6bc9ac36" containerName="keystone-bootstrap" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.135634 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.141554 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d9c876cd8-zfctt" event={"ID":"122ae99a-2891-4bdb-903a-9cccfb3df2a1","Type":"ContainerStarted","Data":"190b3e3dd81572c8823aa4891bba679757b692f8003eef5314dcba4c42923ef6"} Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.141603 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d9c876cd8-zfctt" event={"ID":"122ae99a-2891-4bdb-903a-9cccfb3df2a1","Type":"ContainerStarted","Data":"65af49249f80fee49e4038d4655ea6839dceb13c7bda9011a0a6dcc32fe76550"} Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.141945 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.154009 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.161755 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"06eb71d9-0ccb-4cee-8217-97dae82cbcdf","Type":"ContainerStarted","Data":"8bdbdd7a083cb0978851ed58cc23dffb52dd14e942e37541a37efa80260ebbff"} Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.187703 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sbp9s" event={"ID":"babd629a-b0d5-4ba9-9e01-d36f6bc9ac36","Type":"ContainerDied","Data":"7c7df44aa0a44a3869e0106629043e8211b2d5a842f3852dac77f017def87eb7"} Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.187749 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c7df44aa0a44a3869e0106629043e8211b2d5a842f3852dac77f017def87eb7" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.194519 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-64cc7bf6f4-fcxhm"] Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.195414 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sbp9s" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.298917 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-config-data\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.299030 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-combined-ca-bundle\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.299066 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-public-tls-certs\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.299117 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-credential-keys\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.299156 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-fernet-keys\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.299251 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5h6k\" (UniqueName: \"kubernetes.io/projected/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-kube-api-access-d5h6k\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.299374 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-scripts\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.299425 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-internal-tls-certs\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.404466 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5h6k\" (UniqueName: \"kubernetes.io/projected/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-kube-api-access-d5h6k\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.404527 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-scripts\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.404558 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-internal-tls-certs\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.404600 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-config-data\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.404622 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-combined-ca-bundle\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.404641 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-public-tls-certs\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.404669 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-credential-keys\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.404687 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-fernet-keys\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.415639 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-fernet-keys\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.418258 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-scripts\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.431381 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-credential-keys\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.432563 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-config-data\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.433193 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-internal-tls-certs\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.437736 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-public-tls-certs\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.440907 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5h6k\" (UniqueName: \"kubernetes.io/projected/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-kube-api-access-d5h6k\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.450420 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5-combined-ca-bundle\") pod \"keystone-64cc7bf6f4-fcxhm\" (UID: \"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5\") " pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.511105 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.707855 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.825692 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d855714a-f641-44bc-9b95-ebda879968b0-secret-volume\") pod \"d855714a-f641-44bc-9b95-ebda879968b0\" (UID: \"d855714a-f641-44bc-9b95-ebda879968b0\") " Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.826083 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d855714a-f641-44bc-9b95-ebda879968b0-config-volume\") pod \"d855714a-f641-44bc-9b95-ebda879968b0\" (UID: \"d855714a-f641-44bc-9b95-ebda879968b0\") " Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.826112 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ss7r7\" (UniqueName: \"kubernetes.io/projected/d855714a-f641-44bc-9b95-ebda879968b0-kube-api-access-ss7r7\") pod \"d855714a-f641-44bc-9b95-ebda879968b0\" (UID: \"d855714a-f641-44bc-9b95-ebda879968b0\") " Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.832533 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d855714a-f641-44bc-9b95-ebda879968b0-kube-api-access-ss7r7" (OuterVolumeSpecName: "kube-api-access-ss7r7") pod "d855714a-f641-44bc-9b95-ebda879968b0" (UID: "d855714a-f641-44bc-9b95-ebda879968b0"). InnerVolumeSpecName "kube-api-access-ss7r7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.832992 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d855714a-f641-44bc-9b95-ebda879968b0-config-volume" (OuterVolumeSpecName: "config-volume") pod "d855714a-f641-44bc-9b95-ebda879968b0" (UID: "d855714a-f641-44bc-9b95-ebda879968b0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.837599 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d855714a-f641-44bc-9b95-ebda879968b0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d855714a-f641-44bc-9b95-ebda879968b0" (UID: "d855714a-f641-44bc-9b95-ebda879968b0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.933600 4806 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d855714a-f641-44bc-9b95-ebda879968b0-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.933639 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ss7r7\" (UniqueName: \"kubernetes.io/projected/d855714a-f641-44bc-9b95-ebda879968b0-kube-api-access-ss7r7\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:06 crc kubenswrapper[4806]: I1204 04:15:06.933652 4806 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d855714a-f641-44bc-9b95-ebda879968b0-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:07 crc kubenswrapper[4806]: I1204 04:15:07.246763 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d9c876cd8-zfctt" event={"ID":"122ae99a-2891-4bdb-903a-9cccfb3df2a1","Type":"ContainerStarted","Data":"da6b48ccb1adbab83504a3de8bce5e2cfef6b9e08c330f507d1894a0935fc54d"} Dec 04 04:15:07 crc kubenswrapper[4806]: I1204 04:15:07.247133 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:07 crc kubenswrapper[4806]: I1204 04:15:07.247155 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:07 crc kubenswrapper[4806]: I1204 04:15:07.261019 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" Dec 04 04:15:07 crc kubenswrapper[4806]: I1204 04:15:07.261029 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq" event={"ID":"d855714a-f641-44bc-9b95-ebda879968b0","Type":"ContainerDied","Data":"d20bc5f45571f0bad04fcea30592e462fbddc1ed891a0f6082ea85250ab4a02a"} Dec 04 04:15:07 crc kubenswrapper[4806]: I1204 04:15:07.261089 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d20bc5f45571f0bad04fcea30592e462fbddc1ed891a0f6082ea85250ab4a02a" Dec 04 04:15:07 crc kubenswrapper[4806]: I1204 04:15:07.272989 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-64cc7bf6f4-fcxhm"] Dec 04 04:15:07 crc kubenswrapper[4806]: I1204 04:15:07.296844 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"06eb71d9-0ccb-4cee-8217-97dae82cbcdf","Type":"ContainerStarted","Data":"39f515cae2496194a873198e9bcc3ae7bc4668ff452c27135e6987ebf4276afc"} Dec 04 04:15:07 crc kubenswrapper[4806]: I1204 04:15:07.303458 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-d9c876cd8-zfctt" podStartSLOduration=4.303434325 podStartE2EDuration="4.303434325s" podCreationTimestamp="2025-12-04 04:15:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:15:07.273522026 +0000 UTC m=+1222.132034954" watchObservedRunningTime="2025-12-04 04:15:07.303434325 +0000 UTC m=+1222.161947253" Dec 04 04:15:07 crc kubenswrapper[4806]: I1204 04:15:07.328393 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"60587700-0430-4cc5-a572-234a2a946748","Type":"ContainerStarted","Data":"0ab275c492bba0348963efaa7a72714cbf7dc07e88ae48f2852a33149e3aa1b2"} Dec 04 04:15:07 crc kubenswrapper[4806]: I1204 04:15:07.354865 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-r76nk" event={"ID":"a539706c-9990-4bf7-b393-6fe0bc9f3b0e","Type":"ContainerStarted","Data":"e9c5b59d387e80e532bff56638d010b655f50650443eba16c37f1f4c0bef3721"} Dec 04 04:15:07 crc kubenswrapper[4806]: I1204 04:15:07.753988 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-r76nk" podStartSLOduration=6.288164339 podStartE2EDuration="57.753968832s" podCreationTimestamp="2025-12-04 04:14:10 +0000 UTC" firstStartedPulling="2025-12-04 04:14:12.640118877 +0000 UTC m=+1167.498631795" lastFinishedPulling="2025-12-04 04:15:04.10592336 +0000 UTC m=+1218.964436288" observedRunningTime="2025-12-04 04:15:07.382642032 +0000 UTC m=+1222.241154960" watchObservedRunningTime="2025-12-04 04:15:07.753968832 +0000 UTC m=+1222.612481760" Dec 04 04:15:08 crc kubenswrapper[4806]: I1204 04:15:08.122132 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:15:08 crc kubenswrapper[4806]: I1204 04:15:08.280283 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-rl9g4"] Dec 04 04:15:08 crc kubenswrapper[4806]: I1204 04:15:08.280772 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-rl9g4" podUID="00e618da-f470-4f5e-8955-f5dfd900eceb" containerName="dnsmasq-dns" containerID="cri-o://621825b6b9c53dbec054352c702ef78629d930aed0dabd3cfba3e004932e9d8d" gracePeriod=10 Dec 04 04:15:08 crc kubenswrapper[4806]: I1204 04:15:08.440030 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-64cc7bf6f4-fcxhm" event={"ID":"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5","Type":"ContainerStarted","Data":"fc61e0168b022313f673a56b338513000fbdb1bf6c83061d40ef7127661b4053"} Dec 04 04:15:08 crc kubenswrapper[4806]: I1204 04:15:08.440086 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-64cc7bf6f4-fcxhm" event={"ID":"0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5","Type":"ContainerStarted","Data":"478b7f56cc082d19974024d40ec3d6e14e8c83ffa884d379468450239dee858b"} Dec 04 04:15:08 crc kubenswrapper[4806]: I1204 04:15:08.440144 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:08 crc kubenswrapper[4806]: I1204 04:15:08.473798 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-64cc7bf6f4-fcxhm" podStartSLOduration=2.473775854 podStartE2EDuration="2.473775854s" podCreationTimestamp="2025-12-04 04:15:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:15:08.468559921 +0000 UTC m=+1223.327072849" watchObservedRunningTime="2025-12-04 04:15:08.473775854 +0000 UTC m=+1223.332288782" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.246019 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.322125 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-ovsdbserver-nb\") pod \"00e618da-f470-4f5e-8955-f5dfd900eceb\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.322191 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-config\") pod \"00e618da-f470-4f5e-8955-f5dfd900eceb\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.322225 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-dns-svc\") pod \"00e618da-f470-4f5e-8955-f5dfd900eceb\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.322278 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-ovsdbserver-sb\") pod \"00e618da-f470-4f5e-8955-f5dfd900eceb\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.322299 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2nts\" (UniqueName: \"kubernetes.io/projected/00e618da-f470-4f5e-8955-f5dfd900eceb-kube-api-access-c2nts\") pod \"00e618da-f470-4f5e-8955-f5dfd900eceb\" (UID: \"00e618da-f470-4f5e-8955-f5dfd900eceb\") " Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.350842 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00e618da-f470-4f5e-8955-f5dfd900eceb-kube-api-access-c2nts" (OuterVolumeSpecName: "kube-api-access-c2nts") pod "00e618da-f470-4f5e-8955-f5dfd900eceb" (UID: "00e618da-f470-4f5e-8955-f5dfd900eceb"). InnerVolumeSpecName "kube-api-access-c2nts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.418279 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "00e618da-f470-4f5e-8955-f5dfd900eceb" (UID: "00e618da-f470-4f5e-8955-f5dfd900eceb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.428690 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.428997 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2nts\" (UniqueName: \"kubernetes.io/projected/00e618da-f470-4f5e-8955-f5dfd900eceb-kube-api-access-c2nts\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.536164 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "00e618da-f470-4f5e-8955-f5dfd900eceb" (UID: "00e618da-f470-4f5e-8955-f5dfd900eceb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.537066 4806 generic.go:334] "Generic (PLEG): container finished" podID="00e618da-f470-4f5e-8955-f5dfd900eceb" containerID="621825b6b9c53dbec054352c702ef78629d930aed0dabd3cfba3e004932e9d8d" exitCode=0 Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.537222 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-rl9g4" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.540734 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-rl9g4" event={"ID":"00e618da-f470-4f5e-8955-f5dfd900eceb","Type":"ContainerDied","Data":"621825b6b9c53dbec054352c702ef78629d930aed0dabd3cfba3e004932e9d8d"} Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.540774 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-rl9g4" event={"ID":"00e618da-f470-4f5e-8955-f5dfd900eceb","Type":"ContainerDied","Data":"2e05c4391486237d7e6c376064cb883a0e8f6ab13a14b7acc30be5b959154a03"} Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.540795 4806 scope.go:117] "RemoveContainer" containerID="621825b6b9c53dbec054352c702ef78629d930aed0dabd3cfba3e004932e9d8d" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.570378 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"06eb71d9-0ccb-4cee-8217-97dae82cbcdf","Type":"ContainerStarted","Data":"8428b43f132768c89eff84a87a2ffb8f9182b7bbc7b553fed98c07bff6da1ba9"} Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.574882 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"60587700-0430-4cc5-a572-234a2a946748","Type":"ContainerStarted","Data":"995f3e6bceddd25455c101aa2af9f6e81db185f21e393da766f60b43f9a63f19"} Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.618349 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.618325194 podStartE2EDuration="6.618325194s" podCreationTimestamp="2025-12-04 04:15:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:15:09.606445902 +0000 UTC m=+1224.464958830" watchObservedRunningTime="2025-12-04 04:15:09.618325194 +0000 UTC m=+1224.476838132" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.630436 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "00e618da-f470-4f5e-8955-f5dfd900eceb" (UID: "00e618da-f470-4f5e-8955-f5dfd900eceb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.635347 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.635392 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.642997 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-config" (OuterVolumeSpecName: "config") pod "00e618da-f470-4f5e-8955-f5dfd900eceb" (UID: "00e618da-f470-4f5e-8955-f5dfd900eceb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.651115 4806 scope.go:117] "RemoveContainer" containerID="a16b6f3ed76582a93b36d1f00cac416d552c7de94a4ff28fdf85c0b1a9adbc5c" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.683043 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.683016196 podStartE2EDuration="6.683016196s" podCreationTimestamp="2025-12-04 04:15:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:15:09.653612843 +0000 UTC m=+1224.512125771" watchObservedRunningTime="2025-12-04 04:15:09.683016196 +0000 UTC m=+1224.541529124" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.700324 4806 scope.go:117] "RemoveContainer" containerID="621825b6b9c53dbec054352c702ef78629d930aed0dabd3cfba3e004932e9d8d" Dec 04 04:15:09 crc kubenswrapper[4806]: E1204 04:15:09.700774 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"621825b6b9c53dbec054352c702ef78629d930aed0dabd3cfba3e004932e9d8d\": container with ID starting with 621825b6b9c53dbec054352c702ef78629d930aed0dabd3cfba3e004932e9d8d not found: ID does not exist" containerID="621825b6b9c53dbec054352c702ef78629d930aed0dabd3cfba3e004932e9d8d" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.700811 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"621825b6b9c53dbec054352c702ef78629d930aed0dabd3cfba3e004932e9d8d"} err="failed to get container status \"621825b6b9c53dbec054352c702ef78629d930aed0dabd3cfba3e004932e9d8d\": rpc error: code = NotFound desc = could not find container \"621825b6b9c53dbec054352c702ef78629d930aed0dabd3cfba3e004932e9d8d\": container with ID starting with 621825b6b9c53dbec054352c702ef78629d930aed0dabd3cfba3e004932e9d8d not found: ID does not exist" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.700832 4806 scope.go:117] "RemoveContainer" containerID="a16b6f3ed76582a93b36d1f00cac416d552c7de94a4ff28fdf85c0b1a9adbc5c" Dec 04 04:15:09 crc kubenswrapper[4806]: E1204 04:15:09.701844 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a16b6f3ed76582a93b36d1f00cac416d552c7de94a4ff28fdf85c0b1a9adbc5c\": container with ID starting with a16b6f3ed76582a93b36d1f00cac416d552c7de94a4ff28fdf85c0b1a9adbc5c not found: ID does not exist" containerID="a16b6f3ed76582a93b36d1f00cac416d552c7de94a4ff28fdf85c0b1a9adbc5c" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.701875 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a16b6f3ed76582a93b36d1f00cac416d552c7de94a4ff28fdf85c0b1a9adbc5c"} err="failed to get container status \"a16b6f3ed76582a93b36d1f00cac416d552c7de94a4ff28fdf85c0b1a9adbc5c\": rpc error: code = NotFound desc = could not find container \"a16b6f3ed76582a93b36d1f00cac416d552c7de94a4ff28fdf85c0b1a9adbc5c\": container with ID starting with a16b6f3ed76582a93b36d1f00cac416d552c7de94a4ff28fdf85c0b1a9adbc5c not found: ID does not exist" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.739231 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e618da-f470-4f5e-8955-f5dfd900eceb-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.907003 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-rl9g4"] Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.918828 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-rl9g4"] Dec 04 04:15:09 crc kubenswrapper[4806]: I1204 04:15:09.945438 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bc4f74d8d-xj2xn" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 04 04:15:10 crc kubenswrapper[4806]: I1204 04:15:10.378477 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d8bd6c96d-4sqhf" podUID="d2d53864-c3c8-4e40-8791-0bac49ba561e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 04:15:11 crc kubenswrapper[4806]: I1204 04:15:11.441612 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00e618da-f470-4f5e-8955-f5dfd900eceb" path="/var/lib/kubelet/pods/00e618da-f470-4f5e-8955-f5dfd900eceb/volumes" Dec 04 04:15:12 crc kubenswrapper[4806]: I1204 04:15:12.610632 4806 generic.go:334] "Generic (PLEG): container finished" podID="b1e9cfb9-1741-4a4e-a05d-7d666699876c" containerID="28cf550b5b6b5bf32045b1666c1c389a131dee767915bf17171311afddd538f3" exitCode=0 Dec 04 04:15:12 crc kubenswrapper[4806]: I1204 04:15:12.610753 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pl5fx" event={"ID":"b1e9cfb9-1741-4a4e-a05d-7d666699876c","Type":"ContainerDied","Data":"28cf550b5b6b5bf32045b1666c1c389a131dee767915bf17171311afddd538f3"} Dec 04 04:15:13 crc kubenswrapper[4806]: I1204 04:15:13.803307 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 04:15:13 crc kubenswrapper[4806]: I1204 04:15:13.803706 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 04:15:13 crc kubenswrapper[4806]: I1204 04:15:13.926443 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 04:15:13 crc kubenswrapper[4806]: I1204 04:15:13.983632 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 04:15:14 crc kubenswrapper[4806]: I1204 04:15:14.533172 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 04:15:14 crc kubenswrapper[4806]: I1204 04:15:14.533233 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 04:15:14 crc kubenswrapper[4806]: I1204 04:15:14.581905 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 04:15:14 crc kubenswrapper[4806]: I1204 04:15:14.595778 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 04:15:14 crc kubenswrapper[4806]: I1204 04:15:14.649147 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 04:15:14 crc kubenswrapper[4806]: I1204 04:15:14.649184 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 04:15:14 crc kubenswrapper[4806]: I1204 04:15:14.649194 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 04:15:14 crc kubenswrapper[4806]: I1204 04:15:14.649209 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 04:15:16 crc kubenswrapper[4806]: I1204 04:15:16.676045 4806 generic.go:334] "Generic (PLEG): container finished" podID="a539706c-9990-4bf7-b393-6fe0bc9f3b0e" containerID="e9c5b59d387e80e532bff56638d010b655f50650443eba16c37f1f4c0bef3721" exitCode=0 Dec 04 04:15:16 crc kubenswrapper[4806]: I1204 04:15:16.676128 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-r76nk" event={"ID":"a539706c-9990-4bf7-b393-6fe0bc9f3b0e","Type":"ContainerDied","Data":"e9c5b59d387e80e532bff56638d010b655f50650443eba16c37f1f4c0bef3721"} Dec 04 04:15:18 crc kubenswrapper[4806]: I1204 04:15:18.595568 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 04:15:18 crc kubenswrapper[4806]: I1204 04:15:18.596060 4806 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 04:15:18 crc kubenswrapper[4806]: I1204 04:15:18.612386 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 04:15:18 crc kubenswrapper[4806]: I1204 04:15:18.697819 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 04:15:18 crc kubenswrapper[4806]: I1204 04:15:18.698169 4806 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 04:15:18 crc kubenswrapper[4806]: I1204 04:15:18.700489 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 04:15:19 crc kubenswrapper[4806]: I1204 04:15:19.482766 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pl5fx" Dec 04 04:15:19 crc kubenswrapper[4806]: I1204 04:15:19.648394 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1e9cfb9-1741-4a4e-a05d-7d666699876c-db-sync-config-data\") pod \"b1e9cfb9-1741-4a4e-a05d-7d666699876c\" (UID: \"b1e9cfb9-1741-4a4e-a05d-7d666699876c\") " Dec 04 04:15:19 crc kubenswrapper[4806]: I1204 04:15:19.648456 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e9cfb9-1741-4a4e-a05d-7d666699876c-combined-ca-bundle\") pod \"b1e9cfb9-1741-4a4e-a05d-7d666699876c\" (UID: \"b1e9cfb9-1741-4a4e-a05d-7d666699876c\") " Dec 04 04:15:19 crc kubenswrapper[4806]: I1204 04:15:19.648506 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg2hj\" (UniqueName: \"kubernetes.io/projected/b1e9cfb9-1741-4a4e-a05d-7d666699876c-kube-api-access-mg2hj\") pod \"b1e9cfb9-1741-4a4e-a05d-7d666699876c\" (UID: \"b1e9cfb9-1741-4a4e-a05d-7d666699876c\") " Dec 04 04:15:19 crc kubenswrapper[4806]: I1204 04:15:19.656887 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1e9cfb9-1741-4a4e-a05d-7d666699876c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b1e9cfb9-1741-4a4e-a05d-7d666699876c" (UID: "b1e9cfb9-1741-4a4e-a05d-7d666699876c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:19 crc kubenswrapper[4806]: I1204 04:15:19.672459 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1e9cfb9-1741-4a4e-a05d-7d666699876c-kube-api-access-mg2hj" (OuterVolumeSpecName: "kube-api-access-mg2hj") pod "b1e9cfb9-1741-4a4e-a05d-7d666699876c" (UID: "b1e9cfb9-1741-4a4e-a05d-7d666699876c"). InnerVolumeSpecName "kube-api-access-mg2hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:15:19 crc kubenswrapper[4806]: I1204 04:15:19.684282 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1e9cfb9-1741-4a4e-a05d-7d666699876c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1e9cfb9-1741-4a4e-a05d-7d666699876c" (UID: "b1e9cfb9-1741-4a4e-a05d-7d666699876c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:19 crc kubenswrapper[4806]: I1204 04:15:19.725729 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-pl5fx" Dec 04 04:15:19 crc kubenswrapper[4806]: I1204 04:15:19.727078 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-pl5fx" event={"ID":"b1e9cfb9-1741-4a4e-a05d-7d666699876c","Type":"ContainerDied","Data":"8fe43f7521c70aca171b052f533bd124a8c5cd5af3d8c2cbbca4a11e5adb520a"} Dec 04 04:15:19 crc kubenswrapper[4806]: I1204 04:15:19.727113 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fe43f7521c70aca171b052f533bd124a8c5cd5af3d8c2cbbca4a11e5adb520a" Dec 04 04:15:19 crc kubenswrapper[4806]: I1204 04:15:19.750111 4806 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b1e9cfb9-1741-4a4e-a05d-7d666699876c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:19 crc kubenswrapper[4806]: I1204 04:15:19.750140 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e9cfb9-1741-4a4e-a05d-7d666699876c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:19 crc kubenswrapper[4806]: I1204 04:15:19.750496 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg2hj\" (UniqueName: \"kubernetes.io/projected/b1e9cfb9-1741-4a4e-a05d-7d666699876c-kube-api-access-mg2hj\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:19 crc kubenswrapper[4806]: I1204 04:15:19.944078 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bc4f74d8d-xj2xn" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.380540 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d8bd6c96d-4sqhf" podUID="d2d53864-c3c8-4e40-8791-0bac49ba561e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.777600 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6785cc669c-hgz8x"] Dec 04 04:15:20 crc kubenswrapper[4806]: E1204 04:15:20.777998 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e618da-f470-4f5e-8955-f5dfd900eceb" containerName="dnsmasq-dns" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.778012 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e618da-f470-4f5e-8955-f5dfd900eceb" containerName="dnsmasq-dns" Dec 04 04:15:20 crc kubenswrapper[4806]: E1204 04:15:20.778046 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d855714a-f641-44bc-9b95-ebda879968b0" containerName="collect-profiles" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.778052 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d855714a-f641-44bc-9b95-ebda879968b0" containerName="collect-profiles" Dec 04 04:15:20 crc kubenswrapper[4806]: E1204 04:15:20.778065 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e618da-f470-4f5e-8955-f5dfd900eceb" containerName="init" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.778071 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e618da-f470-4f5e-8955-f5dfd900eceb" containerName="init" Dec 04 04:15:20 crc kubenswrapper[4806]: E1204 04:15:20.778083 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e9cfb9-1741-4a4e-a05d-7d666699876c" containerName="barbican-db-sync" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.778089 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e9cfb9-1741-4a4e-a05d-7d666699876c" containerName="barbican-db-sync" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.778246 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d855714a-f641-44bc-9b95-ebda879968b0" containerName="collect-profiles" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.778262 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="00e618da-f470-4f5e-8955-f5dfd900eceb" containerName="dnsmasq-dns" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.778272 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1e9cfb9-1741-4a4e-a05d-7d666699876c" containerName="barbican-db-sync" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.779220 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.787024 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.789386 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.789601 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-ctxrq" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.804382 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6785cc669c-hgz8x"] Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.872817 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc467d3b-daf0-4869-9cc3-1034cdfd373a-combined-ca-bundle\") pod \"barbican-worker-6785cc669c-hgz8x\" (UID: \"cc467d3b-daf0-4869-9cc3-1034cdfd373a\") " pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.872868 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc467d3b-daf0-4869-9cc3-1034cdfd373a-config-data\") pod \"barbican-worker-6785cc669c-hgz8x\" (UID: \"cc467d3b-daf0-4869-9cc3-1034cdfd373a\") " pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.872890 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc467d3b-daf0-4869-9cc3-1034cdfd373a-config-data-custom\") pod \"barbican-worker-6785cc669c-hgz8x\" (UID: \"cc467d3b-daf0-4869-9cc3-1034cdfd373a\") " pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.872945 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwhdh\" (UniqueName: \"kubernetes.io/projected/cc467d3b-daf0-4869-9cc3-1034cdfd373a-kube-api-access-vwhdh\") pod \"barbican-worker-6785cc669c-hgz8x\" (UID: \"cc467d3b-daf0-4869-9cc3-1034cdfd373a\") " pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.873033 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc467d3b-daf0-4869-9cc3-1034cdfd373a-logs\") pod \"barbican-worker-6785cc669c-hgz8x\" (UID: \"cc467d3b-daf0-4869-9cc3-1034cdfd373a\") " pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.886802 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-b658d66b-ft4r8"] Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.888195 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.892009 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.910257 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-b658d66b-ft4r8"] Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.949355 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-nsfgl"] Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.950718 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.974288 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc467d3b-daf0-4869-9cc3-1034cdfd373a-logs\") pod \"barbican-worker-6785cc669c-hgz8x\" (UID: \"cc467d3b-daf0-4869-9cc3-1034cdfd373a\") " pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.974350 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fef48eb8-ea9d-42ab-8839-fa7a890fe301-config-data-custom\") pod \"barbican-keystone-listener-b658d66b-ft4r8\" (UID: \"fef48eb8-ea9d-42ab-8839-fa7a890fe301\") " pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.974405 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc467d3b-daf0-4869-9cc3-1034cdfd373a-combined-ca-bundle\") pod \"barbican-worker-6785cc669c-hgz8x\" (UID: \"cc467d3b-daf0-4869-9cc3-1034cdfd373a\") " pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.974442 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf85d\" (UniqueName: \"kubernetes.io/projected/fef48eb8-ea9d-42ab-8839-fa7a890fe301-kube-api-access-tf85d\") pod \"barbican-keystone-listener-b658d66b-ft4r8\" (UID: \"fef48eb8-ea9d-42ab-8839-fa7a890fe301\") " pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.974463 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc467d3b-daf0-4869-9cc3-1034cdfd373a-config-data\") pod \"barbican-worker-6785cc669c-hgz8x\" (UID: \"cc467d3b-daf0-4869-9cc3-1034cdfd373a\") " pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.974483 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc467d3b-daf0-4869-9cc3-1034cdfd373a-config-data-custom\") pod \"barbican-worker-6785cc669c-hgz8x\" (UID: \"cc467d3b-daf0-4869-9cc3-1034cdfd373a\") " pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.974540 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwhdh\" (UniqueName: \"kubernetes.io/projected/cc467d3b-daf0-4869-9cc3-1034cdfd373a-kube-api-access-vwhdh\") pod \"barbican-worker-6785cc669c-hgz8x\" (UID: \"cc467d3b-daf0-4869-9cc3-1034cdfd373a\") " pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.974590 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fef48eb8-ea9d-42ab-8839-fa7a890fe301-combined-ca-bundle\") pod \"barbican-keystone-listener-b658d66b-ft4r8\" (UID: \"fef48eb8-ea9d-42ab-8839-fa7a890fe301\") " pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.974621 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fef48eb8-ea9d-42ab-8839-fa7a890fe301-config-data\") pod \"barbican-keystone-listener-b658d66b-ft4r8\" (UID: \"fef48eb8-ea9d-42ab-8839-fa7a890fe301\") " pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.974655 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fef48eb8-ea9d-42ab-8839-fa7a890fe301-logs\") pod \"barbican-keystone-listener-b658d66b-ft4r8\" (UID: \"fef48eb8-ea9d-42ab-8839-fa7a890fe301\") " pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.975095 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc467d3b-daf0-4869-9cc3-1034cdfd373a-logs\") pod \"barbican-worker-6785cc669c-hgz8x\" (UID: \"cc467d3b-daf0-4869-9cc3-1034cdfd373a\") " pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.988215 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc467d3b-daf0-4869-9cc3-1034cdfd373a-config-data\") pod \"barbican-worker-6785cc669c-hgz8x\" (UID: \"cc467d3b-daf0-4869-9cc3-1034cdfd373a\") " pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.989521 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc467d3b-daf0-4869-9cc3-1034cdfd373a-config-data-custom\") pod \"barbican-worker-6785cc669c-hgz8x\" (UID: \"cc467d3b-daf0-4869-9cc3-1034cdfd373a\") " pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:20 crc kubenswrapper[4806]: I1204 04:15:20.993503 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc467d3b-daf0-4869-9cc3-1034cdfd373a-combined-ca-bundle\") pod \"barbican-worker-6785cc669c-hgz8x\" (UID: \"cc467d3b-daf0-4869-9cc3-1034cdfd373a\") " pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.006990 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-nsfgl"] Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.022393 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwhdh\" (UniqueName: \"kubernetes.io/projected/cc467d3b-daf0-4869-9cc3-1034cdfd373a-kube-api-access-vwhdh\") pod \"barbican-worker-6785cc669c-hgz8x\" (UID: \"cc467d3b-daf0-4869-9cc3-1034cdfd373a\") " pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.075990 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf85d\" (UniqueName: \"kubernetes.io/projected/fef48eb8-ea9d-42ab-8839-fa7a890fe301-kube-api-access-tf85d\") pod \"barbican-keystone-listener-b658d66b-ft4r8\" (UID: \"fef48eb8-ea9d-42ab-8839-fa7a890fe301\") " pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.076322 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.076356 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-config\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.076478 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.076602 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fef48eb8-ea9d-42ab-8839-fa7a890fe301-combined-ca-bundle\") pod \"barbican-keystone-listener-b658d66b-ft4r8\" (UID: \"fef48eb8-ea9d-42ab-8839-fa7a890fe301\") " pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.076685 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fef48eb8-ea9d-42ab-8839-fa7a890fe301-config-data\") pod \"barbican-keystone-listener-b658d66b-ft4r8\" (UID: \"fef48eb8-ea9d-42ab-8839-fa7a890fe301\") " pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.076733 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fef48eb8-ea9d-42ab-8839-fa7a890fe301-logs\") pod \"barbican-keystone-listener-b658d66b-ft4r8\" (UID: \"fef48eb8-ea9d-42ab-8839-fa7a890fe301\") " pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.076827 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.076890 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-dns-svc\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.077081 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fef48eb8-ea9d-42ab-8839-fa7a890fe301-config-data-custom\") pod \"barbican-keystone-listener-b658d66b-ft4r8\" (UID: \"fef48eb8-ea9d-42ab-8839-fa7a890fe301\") " pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.077123 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdz29\" (UniqueName: \"kubernetes.io/projected/ad0b3972-6dc5-4635-be2e-e04b65272d8b-kube-api-access-zdz29\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.077305 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fef48eb8-ea9d-42ab-8839-fa7a890fe301-logs\") pod \"barbican-keystone-listener-b658d66b-ft4r8\" (UID: \"fef48eb8-ea9d-42ab-8839-fa7a890fe301\") " pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.096751 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fef48eb8-ea9d-42ab-8839-fa7a890fe301-combined-ca-bundle\") pod \"barbican-keystone-listener-b658d66b-ft4r8\" (UID: \"fef48eb8-ea9d-42ab-8839-fa7a890fe301\") " pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.106575 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fef48eb8-ea9d-42ab-8839-fa7a890fe301-config-data\") pod \"barbican-keystone-listener-b658d66b-ft4r8\" (UID: \"fef48eb8-ea9d-42ab-8839-fa7a890fe301\") " pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.109386 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf85d\" (UniqueName: \"kubernetes.io/projected/fef48eb8-ea9d-42ab-8839-fa7a890fe301-kube-api-access-tf85d\") pod \"barbican-keystone-listener-b658d66b-ft4r8\" (UID: \"fef48eb8-ea9d-42ab-8839-fa7a890fe301\") " pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.122611 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fef48eb8-ea9d-42ab-8839-fa7a890fe301-config-data-custom\") pod \"barbican-keystone-listener-b658d66b-ft4r8\" (UID: \"fef48eb8-ea9d-42ab-8839-fa7a890fe301\") " pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.128636 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6785cc669c-hgz8x" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.162840 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5b8bd497fd-fhbpm"] Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.164734 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.171317 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.178995 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-config\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.179060 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.179158 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.179189 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-dns-svc\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.179214 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdz29\" (UniqueName: \"kubernetes.io/projected/ad0b3972-6dc5-4635-be2e-e04b65272d8b-kube-api-access-zdz29\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.179297 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.180108 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-config\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.180136 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.180832 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.187472 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-dns-svc\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.187804 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.208939 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5b8bd497fd-fhbpm"] Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.216029 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdz29\" (UniqueName: \"kubernetes.io/projected/ad0b3972-6dc5-4635-be2e-e04b65272d8b-kube-api-access-zdz29\") pod \"dnsmasq-dns-85ff748b95-nsfgl\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.224459 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.279257 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.280797 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-config-data-custom\") pod \"barbican-api-5b8bd497fd-fhbpm\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.280933 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c395cfa-4cad-40eb-8362-cb2819fa2024-logs\") pod \"barbican-api-5b8bd497fd-fhbpm\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.280964 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-config-data\") pod \"barbican-api-5b8bd497fd-fhbpm\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.280997 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l84g\" (UniqueName: \"kubernetes.io/projected/1c395cfa-4cad-40eb-8362-cb2819fa2024-kube-api-access-9l84g\") pod \"barbican-api-5b8bd497fd-fhbpm\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.281029 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-combined-ca-bundle\") pod \"barbican-api-5b8bd497fd-fhbpm\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.383102 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c395cfa-4cad-40eb-8362-cb2819fa2024-logs\") pod \"barbican-api-5b8bd497fd-fhbpm\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.383196 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-config-data\") pod \"barbican-api-5b8bd497fd-fhbpm\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.383231 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l84g\" (UniqueName: \"kubernetes.io/projected/1c395cfa-4cad-40eb-8362-cb2819fa2024-kube-api-access-9l84g\") pod \"barbican-api-5b8bd497fd-fhbpm\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.383263 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-combined-ca-bundle\") pod \"barbican-api-5b8bd497fd-fhbpm\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.383313 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-config-data-custom\") pod \"barbican-api-5b8bd497fd-fhbpm\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.388184 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c395cfa-4cad-40eb-8362-cb2819fa2024-logs\") pod \"barbican-api-5b8bd497fd-fhbpm\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.389442 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-config-data-custom\") pod \"barbican-api-5b8bd497fd-fhbpm\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.393063 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-combined-ca-bundle\") pod \"barbican-api-5b8bd497fd-fhbpm\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.397455 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-config-data\") pod \"barbican-api-5b8bd497fd-fhbpm\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.417484 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l84g\" (UniqueName: \"kubernetes.io/projected/1c395cfa-4cad-40eb-8362-cb2819fa2024-kube-api-access-9l84g\") pod \"barbican-api-5b8bd497fd-fhbpm\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.510279 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.815330 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-r76nk" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.898868 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-combined-ca-bundle\") pod \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.898959 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-scripts\") pod \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.899002 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-db-sync-config-data\") pod \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.899028 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-etc-machine-id\") pod \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.899078 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-config-data\") pod \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.899181 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtk77\" (UniqueName: \"kubernetes.io/projected/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-kube-api-access-qtk77\") pod \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\" (UID: \"a539706c-9990-4bf7-b393-6fe0bc9f3b0e\") " Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.911066 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a539706c-9990-4bf7-b393-6fe0bc9f3b0e" (UID: "a539706c-9990-4bf7-b393-6fe0bc9f3b0e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.937253 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-scripts" (OuterVolumeSpecName: "scripts") pod "a539706c-9990-4bf7-b393-6fe0bc9f3b0e" (UID: "a539706c-9990-4bf7-b393-6fe0bc9f3b0e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.937312 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-kube-api-access-qtk77" (OuterVolumeSpecName: "kube-api-access-qtk77") pod "a539706c-9990-4bf7-b393-6fe0bc9f3b0e" (UID: "a539706c-9990-4bf7-b393-6fe0bc9f3b0e"). InnerVolumeSpecName "kube-api-access-qtk77". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:15:21 crc kubenswrapper[4806]: I1204 04:15:21.937650 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a539706c-9990-4bf7-b393-6fe0bc9f3b0e" (UID: "a539706c-9990-4bf7-b393-6fe0bc9f3b0e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:22 crc kubenswrapper[4806]: I1204 04:15:22.001120 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtk77\" (UniqueName: \"kubernetes.io/projected/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-kube-api-access-qtk77\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:22 crc kubenswrapper[4806]: I1204 04:15:22.001150 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:22 crc kubenswrapper[4806]: I1204 04:15:22.001160 4806 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:22 crc kubenswrapper[4806]: I1204 04:15:22.001169 4806 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:22 crc kubenswrapper[4806]: I1204 04:15:22.028088 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a539706c-9990-4bf7-b393-6fe0bc9f3b0e" (UID: "a539706c-9990-4bf7-b393-6fe0bc9f3b0e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:22 crc kubenswrapper[4806]: I1204 04:15:22.028379 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-config-data" (OuterVolumeSpecName: "config-data") pod "a539706c-9990-4bf7-b393-6fe0bc9f3b0e" (UID: "a539706c-9990-4bf7-b393-6fe0bc9f3b0e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:22 crc kubenswrapper[4806]: I1204 04:15:22.102630 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:22 crc kubenswrapper[4806]: I1204 04:15:22.102845 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a539706c-9990-4bf7-b393-6fe0bc9f3b0e-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:22 crc kubenswrapper[4806]: I1204 04:15:22.844796 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-r76nk" event={"ID":"a539706c-9990-4bf7-b393-6fe0bc9f3b0e","Type":"ContainerDied","Data":"7d5376d5c16c194bfdbd9d0c9de7f1c88c14c0cafe1984a1e6bb4e95461993d4"} Dec 04 04:15:22 crc kubenswrapper[4806]: I1204 04:15:22.845160 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d5376d5c16c194bfdbd9d0c9de7f1c88c14c0cafe1984a1e6bb4e95461993d4" Dec 04 04:15:22 crc kubenswrapper[4806]: I1204 04:15:22.845105 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-r76nk" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.209213 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:15:23 crc kubenswrapper[4806]: E1204 04:15:23.236297 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Dec 04 04:15:23 crc kubenswrapper[4806]: E1204 04:15:23.236664 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dwczp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(4783b024-15d6-44a8-bf7f-5ae7f52ce45f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 04:15:23 crc kubenswrapper[4806]: E1204 04:15:23.241056 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="4783b024-15d6-44a8-bf7f-5ae7f52ce45f" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.351333 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 04:15:23 crc kubenswrapper[4806]: E1204 04:15:23.351799 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a539706c-9990-4bf7-b393-6fe0bc9f3b0e" containerName="cinder-db-sync" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.351814 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a539706c-9990-4bf7-b393-6fe0bc9f3b0e" containerName="cinder-db-sync" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.352041 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="a539706c-9990-4bf7-b393-6fe0bc9f3b0e" containerName="cinder-db-sync" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.353103 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.360781 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.360975 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.366377 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-jjhn5" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.376046 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.377152 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.461694 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-nsfgl"] Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.476467 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.476524 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.476551 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-config-data\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.476594 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-scripts\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.476621 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpfxl\" (UniqueName: \"kubernetes.io/projected/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-kube-api-access-lpfxl\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.476660 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.578381 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.578694 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.578731 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-config-data\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.578775 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-scripts\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.578807 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpfxl\" (UniqueName: \"kubernetes.io/projected/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-kube-api-access-lpfxl\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.578833 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.580646 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.592214 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-scripts\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.593119 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.603583 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5fm5h"] Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.606682 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.610810 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-config-data\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.615347 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.642865 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5fm5h"] Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.692161 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpfxl\" (UniqueName: \"kubernetes.io/projected/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-kube-api-access-lpfxl\") pod \"cinder-scheduler-0\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.788415 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.788478 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-config\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.788547 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.788577 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.788601 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgjsz\" (UniqueName: \"kubernetes.io/projected/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-kube-api-access-sgjsz\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.788623 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.830648 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.861900 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4783b024-15d6-44a8-bf7f-5ae7f52ce45f" containerName="ceilometer-notification-agent" containerID="cri-o://aa18177a72fdef2f8f545ffb843dbf7b28b986c608f90a5543e9f0e9d603633f" gracePeriod=30 Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.862125 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4783b024-15d6-44a8-bf7f-5ae7f52ce45f" containerName="sg-core" containerID="cri-o://c12f9444c959f7c0ad73200475b9e50f02ccf11ba247694c417711e06c86f43e" gracePeriod=30 Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.864596 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.866053 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.871336 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.890855 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-config\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.891003 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.891054 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.891242 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgjsz\" (UniqueName: \"kubernetes.io/projected/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-kube-api-access-sgjsz\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.891276 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.891361 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.892810 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.900814 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-config\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.922203 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.922645 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.928666 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.938441 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.975227 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgjsz\" (UniqueName: \"kubernetes.io/projected/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-kube-api-access-sgjsz\") pod \"dnsmasq-dns-5c9776ccc5-5fm5h\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.994687 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.995781 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-scripts\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.995805 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.995827 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhrs7\" (UniqueName: \"kubernetes.io/projected/7a214aba-ebb8-46dd-beb8-441de36c1582-kube-api-access-xhrs7\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.995851 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-config-data\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.995883 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a214aba-ebb8-46dd-beb8-441de36c1582-logs\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.995932 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7a214aba-ebb8-46dd-beb8-441de36c1582-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:23 crc kubenswrapper[4806]: I1204 04:15:23.995994 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-config-data-custom\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.099095 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-config-data-custom\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.099169 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-scripts\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.099196 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.099227 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhrs7\" (UniqueName: \"kubernetes.io/projected/7a214aba-ebb8-46dd-beb8-441de36c1582-kube-api-access-xhrs7\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.099260 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-config-data\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.099307 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a214aba-ebb8-46dd-beb8-441de36c1582-logs\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.099355 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7a214aba-ebb8-46dd-beb8-441de36c1582-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.099483 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7a214aba-ebb8-46dd-beb8-441de36c1582-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.104442 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a214aba-ebb8-46dd-beb8-441de36c1582-logs\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.105823 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-config-data-custom\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.123485 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-scripts\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.125092 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.125721 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-config-data\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.133116 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhrs7\" (UniqueName: \"kubernetes.io/projected/7a214aba-ebb8-46dd-beb8-441de36c1582-kube-api-access-xhrs7\") pod \"cinder-api-0\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " pod="openstack/cinder-api-0" Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.225421 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.454960 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-b658d66b-ft4r8"] Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.467625 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6785cc669c-hgz8x"] Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.630787 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5b8bd497fd-fhbpm"] Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.659513 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-nsfgl"] Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.885433 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.901454 4806 generic.go:334] "Generic (PLEG): container finished" podID="4783b024-15d6-44a8-bf7f-5ae7f52ce45f" containerID="c12f9444c959f7c0ad73200475b9e50f02ccf11ba247694c417711e06c86f43e" exitCode=2 Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.901549 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4783b024-15d6-44a8-bf7f-5ae7f52ce45f","Type":"ContainerDied","Data":"c12f9444c959f7c0ad73200475b9e50f02ccf11ba247694c417711e06c86f43e"} Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.908453 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5fm5h"] Dec 04 04:15:24 crc kubenswrapper[4806]: W1204 04:15:24.909240 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83c1ae2a_6cca_4ba1_8a30_e285d6023ba3.slice/crio-0ef66db12f357783f39b9da9f04e16bb9344d54259932e83ac4010d2e56a21d9 WatchSource:0}: Error finding container 0ef66db12f357783f39b9da9f04e16bb9344d54259932e83ac4010d2e56a21d9: Status 404 returned error can't find the container with id 0ef66db12f357783f39b9da9f04e16bb9344d54259932e83ac4010d2e56a21d9 Dec 04 04:15:24 crc kubenswrapper[4806]: W1204 04:15:24.920328 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9db2f1c9_7c6e_4d73_957c_a2454bfd2ddb.slice/crio-e600f353e6dc01f6f22a9cdca55a2bff4cf0eb8c33da192ddc6b32c3685fafcb WatchSource:0}: Error finding container e600f353e6dc01f6f22a9cdca55a2bff4cf0eb8c33da192ddc6b32c3685fafcb: Status 404 returned error can't find the container with id e600f353e6dc01f6f22a9cdca55a2bff4cf0eb8c33da192ddc6b32c3685fafcb Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.921847 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6785cc669c-hgz8x" event={"ID":"cc467d3b-daf0-4869-9cc3-1034cdfd373a","Type":"ContainerStarted","Data":"cb96d4036100df178a30f493821240e2578ebdff2ae109522a31fa00ce794530"} Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.928900 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" event={"ID":"fef48eb8-ea9d-42ab-8839-fa7a890fe301","Type":"ContainerStarted","Data":"fee53b27ea988cd5fca31ba44322c06516654dd728df1f0e7485bc803df34923"} Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.940748 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b8bd497fd-fhbpm" event={"ID":"1c395cfa-4cad-40eb-8362-cb2819fa2024","Type":"ContainerStarted","Data":"63648aa25b95162a262f332f2c4fb8d34c5271959c7089456a50dfb0e8e5bf64"} Dec 04 04:15:24 crc kubenswrapper[4806]: I1204 04:15:24.970902 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" event={"ID":"ad0b3972-6dc5-4635-be2e-e04b65272d8b","Type":"ContainerStarted","Data":"0aa587a116466470c90b19f01d61ff59fdef498b860532b1c4469cb9eec4f201"} Dec 04 04:15:25 crc kubenswrapper[4806]: I1204 04:15:25.223375 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 04:15:25 crc kubenswrapper[4806]: W1204 04:15:25.306184 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a214aba_ebb8_46dd_beb8_441de36c1582.slice/crio-7092201095e00355da91bcee701eaef27dfe95e0090bfa2a2eb9d47a61c89f48 WatchSource:0}: Error finding container 7092201095e00355da91bcee701eaef27dfe95e0090bfa2a2eb9d47a61c89f48: Status 404 returned error can't find the container with id 7092201095e00355da91bcee701eaef27dfe95e0090bfa2a2eb9d47a61c89f48 Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.010261 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3","Type":"ContainerStarted","Data":"0ef66db12f357783f39b9da9f04e16bb9344d54259932e83ac4010d2e56a21d9"} Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.033440 4806 generic.go:334] "Generic (PLEG): container finished" podID="9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb" containerID="103034db8a4228010d5eba3287cbfba2232e08969f97b8b344d1f9373eec7fa4" exitCode=0 Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.033560 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" event={"ID":"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb","Type":"ContainerDied","Data":"103034db8a4228010d5eba3287cbfba2232e08969f97b8b344d1f9373eec7fa4"} Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.033595 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" event={"ID":"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb","Type":"ContainerStarted","Data":"e600f353e6dc01f6f22a9cdca55a2bff4cf0eb8c33da192ddc6b32c3685fafcb"} Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.087168 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b8bd497fd-fhbpm" event={"ID":"1c395cfa-4cad-40eb-8362-cb2819fa2024","Type":"ContainerStarted","Data":"44c099c1b1e2667cc7a3c3a248b3234946f9dc0e882330b14185ec61de016538"} Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.087224 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b8bd497fd-fhbpm" event={"ID":"1c395cfa-4cad-40eb-8362-cb2819fa2024","Type":"ContainerStarted","Data":"b84ee829c765879ba892935281145d03c4ca17da2bcde8164c8b78105e71d72e"} Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.087435 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.087986 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.102899 4806 generic.go:334] "Generic (PLEG): container finished" podID="ad0b3972-6dc5-4635-be2e-e04b65272d8b" containerID="b7ec1a23fcc5963028a8a3abc42c3c85ec22e3d4fd24e5360b08c2c431134f55" exitCode=0 Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.103211 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" event={"ID":"ad0b3972-6dc5-4635-be2e-e04b65272d8b","Type":"ContainerDied","Data":"b7ec1a23fcc5963028a8a3abc42c3c85ec22e3d4fd24e5360b08c2c431134f55"} Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.110366 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7a214aba-ebb8-46dd-beb8-441de36c1582","Type":"ContainerStarted","Data":"7092201095e00355da91bcee701eaef27dfe95e0090bfa2a2eb9d47a61c89f48"} Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.152489 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5b8bd497fd-fhbpm" podStartSLOduration=5.152464567 podStartE2EDuration="5.152464567s" podCreationTimestamp="2025-12-04 04:15:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:15:26.136029739 +0000 UTC m=+1240.994542677" watchObservedRunningTime="2025-12-04 04:15:26.152464567 +0000 UTC m=+1241.010977495" Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.642009 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.709710 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-ovsdbserver-sb\") pod \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.709823 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdz29\" (UniqueName: \"kubernetes.io/projected/ad0b3972-6dc5-4635-be2e-e04b65272d8b-kube-api-access-zdz29\") pod \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.709854 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-ovsdbserver-nb\") pod \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.709888 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-dns-svc\") pod \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.709992 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-config\") pod \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.710071 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-dns-swift-storage-0\") pod \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\" (UID: \"ad0b3972-6dc5-4635-be2e-e04b65272d8b\") " Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.729213 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad0b3972-6dc5-4635-be2e-e04b65272d8b-kube-api-access-zdz29" (OuterVolumeSpecName: "kube-api-access-zdz29") pod "ad0b3972-6dc5-4635-be2e-e04b65272d8b" (UID: "ad0b3972-6dc5-4635-be2e-e04b65272d8b"). InnerVolumeSpecName "kube-api-access-zdz29". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.751271 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-config" (OuterVolumeSpecName: "config") pod "ad0b3972-6dc5-4635-be2e-e04b65272d8b" (UID: "ad0b3972-6dc5-4635-be2e-e04b65272d8b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.759529 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ad0b3972-6dc5-4635-be2e-e04b65272d8b" (UID: "ad0b3972-6dc5-4635-be2e-e04b65272d8b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.783236 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ad0b3972-6dc5-4635-be2e-e04b65272d8b" (UID: "ad0b3972-6dc5-4635-be2e-e04b65272d8b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.799542 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ad0b3972-6dc5-4635-be2e-e04b65272d8b" (UID: "ad0b3972-6dc5-4635-be2e-e04b65272d8b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.806908 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ad0b3972-6dc5-4635-be2e-e04b65272d8b" (UID: "ad0b3972-6dc5-4635-be2e-e04b65272d8b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.812188 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.812224 4806 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.812237 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.812249 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdz29\" (UniqueName: \"kubernetes.io/projected/ad0b3972-6dc5-4635-be2e-e04b65272d8b-kube-api-access-zdz29\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.812261 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:26 crc kubenswrapper[4806]: I1204 04:15:26.812275 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad0b3972-6dc5-4635-be2e-e04b65272d8b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:27 crc kubenswrapper[4806]: I1204 04:15:27.046782 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:15:27 crc kubenswrapper[4806]: I1204 04:15:27.047123 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:15:27 crc kubenswrapper[4806]: I1204 04:15:27.136800 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" event={"ID":"ad0b3972-6dc5-4635-be2e-e04b65272d8b","Type":"ContainerDied","Data":"0aa587a116466470c90b19f01d61ff59fdef498b860532b1c4469cb9eec4f201"} Dec 04 04:15:27 crc kubenswrapper[4806]: I1204 04:15:27.136847 4806 scope.go:117] "RemoveContainer" containerID="b7ec1a23fcc5963028a8a3abc42c3c85ec22e3d4fd24e5360b08c2c431134f55" Dec 04 04:15:27 crc kubenswrapper[4806]: I1204 04:15:27.137592 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-nsfgl" Dec 04 04:15:27 crc kubenswrapper[4806]: I1204 04:15:27.259839 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-nsfgl"] Dec 04 04:15:27 crc kubenswrapper[4806]: I1204 04:15:27.273060 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-nsfgl"] Dec 04 04:15:27 crc kubenswrapper[4806]: I1204 04:15:27.289563 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-669cc9c8df-p88wk" Dec 04 04:15:27 crc kubenswrapper[4806]: I1204 04:15:27.386793 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7496d77c9d-kjd5v"] Dec 04 04:15:27 crc kubenswrapper[4806]: I1204 04:15:27.401632 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7496d77c9d-kjd5v" podUID="d6850510-af16-4d30-b30c-54a9bd2186d3" containerName="neutron-api" containerID="cri-o://96112656a08459e761509553df8027474648735003ea0089e07404ec72b6f05c" gracePeriod=30 Dec 04 04:15:27 crc kubenswrapper[4806]: I1204 04:15:27.402199 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7496d77c9d-kjd5v" podUID="d6850510-af16-4d30-b30c-54a9bd2186d3" containerName="neutron-httpd" containerID="cri-o://95dd0182bcbefacae5d7c5b8075c03f37fcb2e14fb34b4ee557b3091755a4825" gracePeriod=30 Dec 04 04:15:27 crc kubenswrapper[4806]: I1204 04:15:27.468122 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad0b3972-6dc5-4635-be2e-e04b65272d8b" path="/var/lib/kubelet/pods/ad0b3972-6dc5-4635-be2e-e04b65272d8b/volumes" Dec 04 04:15:27 crc kubenswrapper[4806]: I1204 04:15:27.533598 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.173527 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" event={"ID":"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb","Type":"ContainerStarted","Data":"bb221910e129a7ccea0e27f287ad0050a8d2c0dfe3ba999e684e38eadf5d055e"} Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.173823 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.196509 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7a214aba-ebb8-46dd-beb8-441de36c1582","Type":"ContainerStarted","Data":"d358de5198367486a96f69e0bcb7e451eee35b4ed1593958036094daeb312232"} Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.211122 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" podStartSLOduration=5.211106719 podStartE2EDuration="5.211106719s" podCreationTimestamp="2025-12-04 04:15:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:15:28.210652235 +0000 UTC m=+1243.069165163" watchObservedRunningTime="2025-12-04 04:15:28.211106719 +0000 UTC m=+1243.069619647" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.213528 4806 generic.go:334] "Generic (PLEG): container finished" podID="d6850510-af16-4d30-b30c-54a9bd2186d3" containerID="95dd0182bcbefacae5d7c5b8075c03f37fcb2e14fb34b4ee557b3091755a4825" exitCode=0 Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.213623 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7496d77c9d-kjd5v" event={"ID":"d6850510-af16-4d30-b30c-54a9bd2186d3","Type":"ContainerDied","Data":"95dd0182bcbefacae5d7c5b8075c03f37fcb2e14fb34b4ee557b3091755a4825"} Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.226672 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3","Type":"ContainerStarted","Data":"c7a92ed6ce652bbe78c097b6b9ee4658d9058abf96a0dcdded3a64674aa24633"} Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.302888 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-949b6795d-z7dx2"] Dec 04 04:15:28 crc kubenswrapper[4806]: E1204 04:15:28.303304 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad0b3972-6dc5-4635-be2e-e04b65272d8b" containerName="init" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.303324 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad0b3972-6dc5-4635-be2e-e04b65272d8b" containerName="init" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.303508 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad0b3972-6dc5-4635-be2e-e04b65272d8b" containerName="init" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.306447 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.315671 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.315917 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.334393 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-949b6795d-z7dx2"] Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.378130 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef61aa0b-2b2c-481f-b22d-4ea770a91711-combined-ca-bundle\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.378431 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6d9j\" (UniqueName: \"kubernetes.io/projected/ef61aa0b-2b2c-481f-b22d-4ea770a91711-kube-api-access-r6d9j\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.380980 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef61aa0b-2b2c-481f-b22d-4ea770a91711-logs\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.381141 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef61aa0b-2b2c-481f-b22d-4ea770a91711-public-tls-certs\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.382209 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef61aa0b-2b2c-481f-b22d-4ea770a91711-internal-tls-certs\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.382793 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef61aa0b-2b2c-481f-b22d-4ea770a91711-config-data-custom\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.383133 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef61aa0b-2b2c-481f-b22d-4ea770a91711-config-data\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.485192 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6d9j\" (UniqueName: \"kubernetes.io/projected/ef61aa0b-2b2c-481f-b22d-4ea770a91711-kube-api-access-r6d9j\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.485256 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef61aa0b-2b2c-481f-b22d-4ea770a91711-logs\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.485283 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef61aa0b-2b2c-481f-b22d-4ea770a91711-public-tls-certs\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.485307 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef61aa0b-2b2c-481f-b22d-4ea770a91711-internal-tls-certs\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.485351 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef61aa0b-2b2c-481f-b22d-4ea770a91711-config-data-custom\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.485467 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef61aa0b-2b2c-481f-b22d-4ea770a91711-config-data\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.485536 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef61aa0b-2b2c-481f-b22d-4ea770a91711-combined-ca-bundle\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.487547 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef61aa0b-2b2c-481f-b22d-4ea770a91711-logs\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.497878 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef61aa0b-2b2c-481f-b22d-4ea770a91711-internal-tls-certs\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.512400 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6d9j\" (UniqueName: \"kubernetes.io/projected/ef61aa0b-2b2c-481f-b22d-4ea770a91711-kube-api-access-r6d9j\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.513184 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef61aa0b-2b2c-481f-b22d-4ea770a91711-config-data\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.520483 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef61aa0b-2b2c-481f-b22d-4ea770a91711-public-tls-certs\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.522678 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef61aa0b-2b2c-481f-b22d-4ea770a91711-combined-ca-bundle\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.523581 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef61aa0b-2b2c-481f-b22d-4ea770a91711-config-data-custom\") pod \"barbican-api-949b6795d-z7dx2\" (UID: \"ef61aa0b-2b2c-481f-b22d-4ea770a91711\") " pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:28 crc kubenswrapper[4806]: I1204 04:15:28.636559 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:29 crc kubenswrapper[4806]: I1204 04:15:29.251485 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7a214aba-ebb8-46dd-beb8-441de36c1582","Type":"ContainerStarted","Data":"fcbd5a372edce403b131f444c2ed98c5ef52c8dc46ede9b087a5f3737f1d05c1"} Dec 04 04:15:29 crc kubenswrapper[4806]: I1204 04:15:29.252195 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7a214aba-ebb8-46dd-beb8-441de36c1582" containerName="cinder-api-log" containerID="cri-o://d358de5198367486a96f69e0bcb7e451eee35b4ed1593958036094daeb312232" gracePeriod=30 Dec 04 04:15:29 crc kubenswrapper[4806]: I1204 04:15:29.252333 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 04 04:15:29 crc kubenswrapper[4806]: I1204 04:15:29.252712 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7a214aba-ebb8-46dd-beb8-441de36c1582" containerName="cinder-api" containerID="cri-o://fcbd5a372edce403b131f444c2ed98c5ef52c8dc46ede9b087a5f3737f1d05c1" gracePeriod=30 Dec 04 04:15:29 crc kubenswrapper[4806]: I1204 04:15:29.274944 4806 generic.go:334] "Generic (PLEG): container finished" podID="4783b024-15d6-44a8-bf7f-5ae7f52ce45f" containerID="aa18177a72fdef2f8f545ffb843dbf7b28b986c608f90a5543e9f0e9d603633f" exitCode=0 Dec 04 04:15:29 crc kubenswrapper[4806]: I1204 04:15:29.275298 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4783b024-15d6-44a8-bf7f-5ae7f52ce45f","Type":"ContainerDied","Data":"aa18177a72fdef2f8f545ffb843dbf7b28b986c608f90a5543e9f0e9d603633f"} Dec 04 04:15:29 crc kubenswrapper[4806]: I1204 04:15:29.315857 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.31583712 podStartE2EDuration="6.31583712s" podCreationTimestamp="2025-12-04 04:15:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:15:29.292422601 +0000 UTC m=+1244.150935529" watchObservedRunningTime="2025-12-04 04:15:29.31583712 +0000 UTC m=+1244.174350048" Dec 04 04:15:29 crc kubenswrapper[4806]: I1204 04:15:29.944409 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bc4f74d8d-xj2xn" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 04 04:15:29 crc kubenswrapper[4806]: I1204 04:15:29.946612 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:15:29 crc kubenswrapper[4806]: I1204 04:15:29.951823 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"33669b61f9a2e337d989305386e72fc8b569e4ce6dbdb13e37c0aacc3aaf1ca1"} pod="openstack/horizon-7bc4f74d8d-xj2xn" containerMessage="Container horizon failed startup probe, will be restarted" Dec 04 04:15:29 crc kubenswrapper[4806]: I1204 04:15:29.952427 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bc4f74d8d-xj2xn" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" containerID="cri-o://33669b61f9a2e337d989305386e72fc8b569e4ce6dbdb13e37c0aacc3aaf1ca1" gracePeriod=30 Dec 04 04:15:29 crc kubenswrapper[4806]: I1204 04:15:29.961138 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:15:29 crc kubenswrapper[4806]: I1204 04:15:29.967076 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-949b6795d-z7dx2"] Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.072395 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-sg-core-conf-yaml\") pod \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.072465 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-run-httpd\") pod \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.072493 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-log-httpd\") pod \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.072608 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-config-data\") pod \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.072642 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-combined-ca-bundle\") pod \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.072697 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwczp\" (UniqueName: \"kubernetes.io/projected/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-kube-api-access-dwczp\") pod \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.072712 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-scripts\") pod \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\" (UID: \"4783b024-15d6-44a8-bf7f-5ae7f52ce45f\") " Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.092642 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4783b024-15d6-44a8-bf7f-5ae7f52ce45f" (UID: "4783b024-15d6-44a8-bf7f-5ae7f52ce45f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.092869 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4783b024-15d6-44a8-bf7f-5ae7f52ce45f" (UID: "4783b024-15d6-44a8-bf7f-5ae7f52ce45f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.114631 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-kube-api-access-dwczp" (OuterVolumeSpecName: "kube-api-access-dwczp") pod "4783b024-15d6-44a8-bf7f-5ae7f52ce45f" (UID: "4783b024-15d6-44a8-bf7f-5ae7f52ce45f"). InnerVolumeSpecName "kube-api-access-dwczp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.127932 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-scripts" (OuterVolumeSpecName: "scripts") pod "4783b024-15d6-44a8-bf7f-5ae7f52ce45f" (UID: "4783b024-15d6-44a8-bf7f-5ae7f52ce45f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.128267 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4783b024-15d6-44a8-bf7f-5ae7f52ce45f" (UID: "4783b024-15d6-44a8-bf7f-5ae7f52ce45f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.150176 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4783b024-15d6-44a8-bf7f-5ae7f52ce45f" (UID: "4783b024-15d6-44a8-bf7f-5ae7f52ce45f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.175381 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.175427 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwczp\" (UniqueName: \"kubernetes.io/projected/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-kube-api-access-dwczp\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.175443 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.175453 4806 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.175468 4806 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.175479 4806 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.211840 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-config-data" (OuterVolumeSpecName: "config-data") pod "4783b024-15d6-44a8-bf7f-5ae7f52ce45f" (UID: "4783b024-15d6-44a8-bf7f-5ae7f52ce45f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.277315 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4783b024-15d6-44a8-bf7f-5ae7f52ce45f-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.302326 4806 generic.go:334] "Generic (PLEG): container finished" podID="7a214aba-ebb8-46dd-beb8-441de36c1582" containerID="d358de5198367486a96f69e0bcb7e451eee35b4ed1593958036094daeb312232" exitCode=143 Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.302429 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7a214aba-ebb8-46dd-beb8-441de36c1582","Type":"ContainerDied","Data":"d358de5198367486a96f69e0bcb7e451eee35b4ed1593958036094daeb312232"} Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.314363 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-949b6795d-z7dx2" event={"ID":"ef61aa0b-2b2c-481f-b22d-4ea770a91711","Type":"ContainerStarted","Data":"7e0bf325499d8a11473d1e570b8d4cc9b9cfb376cf4ead77edb5de88288a78be"} Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.314405 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-949b6795d-z7dx2" event={"ID":"ef61aa0b-2b2c-481f-b22d-4ea770a91711","Type":"ContainerStarted","Data":"b34fddcff2c2ed1cd5a84a80e1027d0838140ac707379ada442d04f660480cd7"} Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.317568 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4783b024-15d6-44a8-bf7f-5ae7f52ce45f","Type":"ContainerDied","Data":"206890fc3d206e0496ba8b02ef7455e9d71e3e90c948cd654e3eb23c4c8c3f19"} Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.317615 4806 scope.go:117] "RemoveContainer" containerID="c12f9444c959f7c0ad73200475b9e50f02ccf11ba247694c417711e06c86f43e" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.317734 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.331110 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6785cc669c-hgz8x" event={"ID":"cc467d3b-daf0-4869-9cc3-1034cdfd373a","Type":"ContainerStarted","Data":"cea7355c9835cc40c549d0a70ca5e5549a5ff031e9383de291db18583382e05e"} Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.342249 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" event={"ID":"fef48eb8-ea9d-42ab-8839-fa7a890fe301","Type":"ContainerStarted","Data":"e653676528d66cf7609dcf99a1c316f96388e928aa6dd66ed054b3f3d74ce65a"} Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.377129 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d8bd6c96d-4sqhf" podUID="d2d53864-c3c8-4e40-8791-0bac49ba561e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.377197 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.377853 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"a440aa262b334c36c507fc4d0fece95b286074ca6be74690b931e0ac6569d085"} pod="openstack/horizon-7d8bd6c96d-4sqhf" containerMessage="Container horizon failed startup probe, will be restarted" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.377879 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d8bd6c96d-4sqhf" podUID="d2d53864-c3c8-4e40-8791-0bac49ba561e" containerName="horizon" containerID="cri-o://a440aa262b334c36c507fc4d0fece95b286074ca6be74690b931e0ac6569d085" gracePeriod=30 Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.412075 4806 scope.go:117] "RemoveContainer" containerID="aa18177a72fdef2f8f545ffb843dbf7b28b986c608f90a5543e9f0e9d603633f" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.469999 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.504001 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.517041 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:15:30 crc kubenswrapper[4806]: E1204 04:15:30.517477 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4783b024-15d6-44a8-bf7f-5ae7f52ce45f" containerName="ceilometer-notification-agent" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.517488 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4783b024-15d6-44a8-bf7f-5ae7f52ce45f" containerName="ceilometer-notification-agent" Dec 04 04:15:30 crc kubenswrapper[4806]: E1204 04:15:30.517509 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4783b024-15d6-44a8-bf7f-5ae7f52ce45f" containerName="sg-core" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.517516 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4783b024-15d6-44a8-bf7f-5ae7f52ce45f" containerName="sg-core" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.517676 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="4783b024-15d6-44a8-bf7f-5ae7f52ce45f" containerName="ceilometer-notification-agent" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.517704 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="4783b024-15d6-44a8-bf7f-5ae7f52ce45f" containerName="sg-core" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.519275 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.523535 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.523684 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.543515 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.582364 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-scripts\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.582461 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.582480 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-run-httpd\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.582531 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd7kr\" (UniqueName: \"kubernetes.io/projected/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-kube-api-access-cd7kr\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.582568 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-log-httpd\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.582600 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.582630 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-config-data\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.687573 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-log-httpd\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.687635 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.687670 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-config-data\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.687699 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-scripts\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.687797 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.687817 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-run-httpd\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.687952 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd7kr\" (UniqueName: \"kubernetes.io/projected/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-kube-api-access-cd7kr\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.688717 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-log-httpd\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.691570 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-run-httpd\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.696731 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-scripts\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.697255 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.698656 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-config-data\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.699160 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.716821 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd7kr\" (UniqueName: \"kubernetes.io/projected/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-kube-api-access-cd7kr\") pod \"ceilometer-0\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " pod="openstack/ceilometer-0" Dec 04 04:15:30 crc kubenswrapper[4806]: I1204 04:15:30.838386 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.457100 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4783b024-15d6-44a8-bf7f-5ae7f52ce45f" path="/var/lib/kubelet/pods/4783b024-15d6-44a8-bf7f-5ae7f52ce45f/volumes" Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.458452 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3","Type":"ContainerStarted","Data":"c6109f8c356b8024f08644d074a1b5e27920bac999614a1ccac2d89260cc15d6"} Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.463323 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6785cc669c-hgz8x" event={"ID":"cc467d3b-daf0-4869-9cc3-1034cdfd373a","Type":"ContainerStarted","Data":"efa39d6f55880fd2abae6f2350587bee68cbdb0e316f35b15942b33081c458a3"} Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.470376 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" event={"ID":"fef48eb8-ea9d-42ab-8839-fa7a890fe301","Type":"ContainerStarted","Data":"876c383524c7e7b7f6d6c9494497d06fe912a63340a4c4a4638bdc317ca5638c"} Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.499305 4806 generic.go:334] "Generic (PLEG): container finished" podID="7a214aba-ebb8-46dd-beb8-441de36c1582" containerID="fcbd5a372edce403b131f444c2ed98c5ef52c8dc46ede9b087a5f3737f1d05c1" exitCode=0 Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.499387 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7a214aba-ebb8-46dd-beb8-441de36c1582","Type":"ContainerDied","Data":"fcbd5a372edce403b131f444c2ed98c5ef52c8dc46ede9b087a5f3737f1d05c1"} Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.517320 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-949b6795d-z7dx2" event={"ID":"ef61aa0b-2b2c-481f-b22d-4ea770a91711","Type":"ContainerStarted","Data":"b152a8ac9a27896310bd33c5ec0a18d973e9c12489236bcef719cd2f67413e32"} Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.518272 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.518305 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.529404 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=7.057178054 podStartE2EDuration="8.52937615s" podCreationTimestamp="2025-12-04 04:15:23 +0000 UTC" firstStartedPulling="2025-12-04 04:15:24.922101713 +0000 UTC m=+1239.780614641" lastFinishedPulling="2025-12-04 04:15:26.394299809 +0000 UTC m=+1241.252812737" observedRunningTime="2025-12-04 04:15:31.469702496 +0000 UTC m=+1246.328215434" watchObservedRunningTime="2025-12-04 04:15:31.52937615 +0000 UTC m=+1246.387889078" Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.598724 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6785cc669c-hgz8x" podStartSLOduration=6.786058911 podStartE2EDuration="11.598704157s" podCreationTimestamp="2025-12-04 04:15:20 +0000 UTC" firstStartedPulling="2025-12-04 04:15:24.5312729 +0000 UTC m=+1239.389785828" lastFinishedPulling="2025-12-04 04:15:29.343918146 +0000 UTC m=+1244.202431074" observedRunningTime="2025-12-04 04:15:31.505432783 +0000 UTC m=+1246.363945731" watchObservedRunningTime="2025-12-04 04:15:31.598704157 +0000 UTC m=+1246.457217075" Dec 04 04:15:31 crc kubenswrapper[4806]: W1204 04:15:31.698066 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7003c3f_33df_4ef5_b58f_83aa0b80cf3a.slice/crio-b613d6faf3ea89efa3ad934817382dc7be856928ea65fdbe14fcc18301077440 WatchSource:0}: Error finding container b613d6faf3ea89efa3ad934817382dc7be856928ea65fdbe14fcc18301077440: Status 404 returned error can't find the container with id b613d6faf3ea89efa3ad934817382dc7be856928ea65fdbe14fcc18301077440 Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.712675 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-b658d66b-ft4r8" podStartSLOduration=6.934886457 podStartE2EDuration="11.712653772s" podCreationTimestamp="2025-12-04 04:15:20 +0000 UTC" firstStartedPulling="2025-12-04 04:15:24.559328236 +0000 UTC m=+1239.417841164" lastFinishedPulling="2025-12-04 04:15:29.337095551 +0000 UTC m=+1244.195608479" observedRunningTime="2025-12-04 04:15:31.576549738 +0000 UTC m=+1246.435062666" watchObservedRunningTime="2025-12-04 04:15:31.712653772 +0000 UTC m=+1246.571166690" Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.760303 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-949b6795d-z7dx2" podStartSLOduration=3.760283696 podStartE2EDuration="3.760283696s" podCreationTimestamp="2025-12-04 04:15:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:15:31.621329791 +0000 UTC m=+1246.479842719" watchObservedRunningTime="2025-12-04 04:15:31.760283696 +0000 UTC m=+1246.618796624" Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.791346 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.798445 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.921543 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-config-data\") pod \"7a214aba-ebb8-46dd-beb8-441de36c1582\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.921615 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-combined-ca-bundle\") pod \"7a214aba-ebb8-46dd-beb8-441de36c1582\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.921639 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a214aba-ebb8-46dd-beb8-441de36c1582-logs\") pod \"7a214aba-ebb8-46dd-beb8-441de36c1582\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.921687 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhrs7\" (UniqueName: \"kubernetes.io/projected/7a214aba-ebb8-46dd-beb8-441de36c1582-kube-api-access-xhrs7\") pod \"7a214aba-ebb8-46dd-beb8-441de36c1582\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.921757 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7a214aba-ebb8-46dd-beb8-441de36c1582-etc-machine-id\") pod \"7a214aba-ebb8-46dd-beb8-441de36c1582\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.921829 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-scripts\") pod \"7a214aba-ebb8-46dd-beb8-441de36c1582\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.921890 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-config-data-custom\") pod \"7a214aba-ebb8-46dd-beb8-441de36c1582\" (UID: \"7a214aba-ebb8-46dd-beb8-441de36c1582\") " Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.922371 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a214aba-ebb8-46dd-beb8-441de36c1582-logs" (OuterVolumeSpecName: "logs") pod "7a214aba-ebb8-46dd-beb8-441de36c1582" (UID: "7a214aba-ebb8-46dd-beb8-441de36c1582"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.922909 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a214aba-ebb8-46dd-beb8-441de36c1582-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7a214aba-ebb8-46dd-beb8-441de36c1582" (UID: "7a214aba-ebb8-46dd-beb8-441de36c1582"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.936571 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7a214aba-ebb8-46dd-beb8-441de36c1582" (UID: "7a214aba-ebb8-46dd-beb8-441de36c1582"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.952204 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-scripts" (OuterVolumeSpecName: "scripts") pod "7a214aba-ebb8-46dd-beb8-441de36c1582" (UID: "7a214aba-ebb8-46dd-beb8-441de36c1582"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:31 crc kubenswrapper[4806]: I1204 04:15:31.974282 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a214aba-ebb8-46dd-beb8-441de36c1582-kube-api-access-xhrs7" (OuterVolumeSpecName: "kube-api-access-xhrs7") pod "7a214aba-ebb8-46dd-beb8-441de36c1582" (UID: "7a214aba-ebb8-46dd-beb8-441de36c1582"). InnerVolumeSpecName "kube-api-access-xhrs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.007199 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-config-data" (OuterVolumeSpecName: "config-data") pod "7a214aba-ebb8-46dd-beb8-441de36c1582" (UID: "7a214aba-ebb8-46dd-beb8-441de36c1582"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.007237 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a214aba-ebb8-46dd-beb8-441de36c1582" (UID: "7a214aba-ebb8-46dd-beb8-441de36c1582"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.024388 4806 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.024433 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.024447 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.024458 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a214aba-ebb8-46dd-beb8-441de36c1582-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.024470 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhrs7\" (UniqueName: \"kubernetes.io/projected/7a214aba-ebb8-46dd-beb8-441de36c1582-kube-api-access-xhrs7\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.024482 4806 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7a214aba-ebb8-46dd-beb8-441de36c1582-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.024489 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a214aba-ebb8-46dd-beb8-441de36c1582-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.577973 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a","Type":"ContainerStarted","Data":"60c3830bf09243bd142f2e2e096840ef7ccc4dc359af29a53725a6282a2334e7"} Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.578314 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a","Type":"ContainerStarted","Data":"b613d6faf3ea89efa3ad934817382dc7be856928ea65fdbe14fcc18301077440"} Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.580899 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7a214aba-ebb8-46dd-beb8-441de36c1582","Type":"ContainerDied","Data":"7092201095e00355da91bcee701eaef27dfe95e0090bfa2a2eb9d47a61c89f48"} Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.581059 4806 scope.go:117] "RemoveContainer" containerID="fcbd5a372edce403b131f444c2ed98c5ef52c8dc46ede9b087a5f3737f1d05c1" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.581245 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.629933 4806 scope.go:117] "RemoveContainer" containerID="d358de5198367486a96f69e0bcb7e451eee35b4ed1593958036094daeb312232" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.673598 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.731321 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.790499 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 04 04:15:32 crc kubenswrapper[4806]: E1204 04:15:32.792821 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a214aba-ebb8-46dd-beb8-441de36c1582" containerName="cinder-api-log" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.793162 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a214aba-ebb8-46dd-beb8-441de36c1582" containerName="cinder-api-log" Dec 04 04:15:32 crc kubenswrapper[4806]: E1204 04:15:32.793312 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a214aba-ebb8-46dd-beb8-441de36c1582" containerName="cinder-api" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.793402 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a214aba-ebb8-46dd-beb8-441de36c1582" containerName="cinder-api" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.794533 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a214aba-ebb8-46dd-beb8-441de36c1582" containerName="cinder-api-log" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.794718 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a214aba-ebb8-46dd-beb8-441de36c1582" containerName="cinder-api" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.800051 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.806523 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.813000 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.815800 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.824350 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.881467 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.882833 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-config-data\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.883018 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d05a5b8-2294-4d30-a0b7-d865c8aae646-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.883181 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcccb\" (UniqueName: \"kubernetes.io/projected/3d05a5b8-2294-4d30-a0b7-d865c8aae646-kube-api-access-hcccb\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.883319 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.883497 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.883623 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-scripts\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.884046 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-config-data-custom\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.884114 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d05a5b8-2294-4d30-a0b7-d865c8aae646-logs\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.986060 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d05a5b8-2294-4d30-a0b7-d865c8aae646-logs\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.986483 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.986536 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-config-data\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.986561 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d05a5b8-2294-4d30-a0b7-d865c8aae646-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.986597 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcccb\" (UniqueName: \"kubernetes.io/projected/3d05a5b8-2294-4d30-a0b7-d865c8aae646-kube-api-access-hcccb\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.986628 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.986684 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.986719 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-scripts\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.986803 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-config-data-custom\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.987368 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d05a5b8-2294-4d30-a0b7-d865c8aae646-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:32 crc kubenswrapper[4806]: I1204 04:15:32.988649 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d05a5b8-2294-4d30-a0b7-d865c8aae646-logs\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.004241 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-config-data-custom\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.004687 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.004725 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.007238 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.023121 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-config-data\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.024813 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d05a5b8-2294-4d30-a0b7-d865c8aae646-scripts\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.025185 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcccb\" (UniqueName: \"kubernetes.io/projected/3d05a5b8-2294-4d30-a0b7-d865c8aae646-kube-api-access-hcccb\") pod \"cinder-api-0\" (UID: \"3d05a5b8-2294-4d30-a0b7-d865c8aae646\") " pod="openstack/cinder-api-0" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.144530 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.472414 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a214aba-ebb8-46dd-beb8-441de36c1582" path="/var/lib/kubelet/pods/7a214aba-ebb8-46dd-beb8-441de36c1582/volumes" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.652154 4806 generic.go:334] "Generic (PLEG): container finished" podID="d6850510-af16-4d30-b30c-54a9bd2186d3" containerID="96112656a08459e761509553df8027474648735003ea0089e07404ec72b6f05c" exitCode=0 Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.652260 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7496d77c9d-kjd5v" event={"ID":"d6850510-af16-4d30-b30c-54a9bd2186d3","Type":"ContainerDied","Data":"96112656a08459e761509553df8027474648735003ea0089e07404ec72b6f05c"} Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.652287 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7496d77c9d-kjd5v" event={"ID":"d6850510-af16-4d30-b30c-54a9bd2186d3","Type":"ContainerDied","Data":"7db00b93a79ccf42a6471519ff561df1a65bd2853a28e50441a06a1284641f08"} Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.652297 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7db00b93a79ccf42a6471519ff561df1a65bd2853a28e50441a06a1284641f08" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.734003 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.821568 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-ovndb-tls-certs\") pod \"d6850510-af16-4d30-b30c-54a9bd2186d3\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.821986 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-combined-ca-bundle\") pod \"d6850510-af16-4d30-b30c-54a9bd2186d3\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.822021 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-httpd-config\") pod \"d6850510-af16-4d30-b30c-54a9bd2186d3\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.822047 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-config\") pod \"d6850510-af16-4d30-b30c-54a9bd2186d3\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.822123 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fs8gf\" (UniqueName: \"kubernetes.io/projected/d6850510-af16-4d30-b30c-54a9bd2186d3-kube-api-access-fs8gf\") pod \"d6850510-af16-4d30-b30c-54a9bd2186d3\" (UID: \"d6850510-af16-4d30-b30c-54a9bd2186d3\") " Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.836178 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.846996 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.160:8080/\": dial tcp 10.217.0.160:8080: connect: connection refused" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.852865 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6850510-af16-4d30-b30c-54a9bd2186d3-kube-api-access-fs8gf" (OuterVolumeSpecName: "kube-api-access-fs8gf") pod "d6850510-af16-4d30-b30c-54a9bd2186d3" (UID: "d6850510-af16-4d30-b30c-54a9bd2186d3"). InnerVolumeSpecName "kube-api-access-fs8gf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.855219 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "d6850510-af16-4d30-b30c-54a9bd2186d3" (UID: "d6850510-af16-4d30-b30c-54a9bd2186d3"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.927379 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fs8gf\" (UniqueName: \"kubernetes.io/projected/d6850510-af16-4d30-b30c-54a9bd2186d3-kube-api-access-fs8gf\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.927629 4806 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.931526 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.977507 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-config" (OuterVolumeSpecName: "config") pod "d6850510-af16-4d30-b30c-54a9bd2186d3" (UID: "d6850510-af16-4d30-b30c-54a9bd2186d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:33 crc kubenswrapper[4806]: I1204 04:15:33.998200 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:15:34 crc kubenswrapper[4806]: I1204 04:15:34.001016 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6850510-af16-4d30-b30c-54a9bd2186d3" (UID: "d6850510-af16-4d30-b30c-54a9bd2186d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:34 crc kubenswrapper[4806]: I1204 04:15:34.033319 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:34 crc kubenswrapper[4806]: I1204 04:15:34.033351 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:34 crc kubenswrapper[4806]: I1204 04:15:34.083172 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "d6850510-af16-4d30-b30c-54a9bd2186d3" (UID: "d6850510-af16-4d30-b30c-54a9bd2186d3"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:34 crc kubenswrapper[4806]: I1204 04:15:34.115629 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-wbj9t"] Dec 04 04:15:34 crc kubenswrapper[4806]: I1204 04:15:34.115869 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" podUID="88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b" containerName="dnsmasq-dns" containerID="cri-o://f5963e6d29ec3803ccbf9650a1ebe9d814b2db349acc32fccccdc1b1278f6bbe" gracePeriod=10 Dec 04 04:15:34 crc kubenswrapper[4806]: I1204 04:15:34.144600 4806 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6850510-af16-4d30-b30c-54a9bd2186d3-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:34 crc kubenswrapper[4806]: I1204 04:15:34.731326 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a","Type":"ContainerStarted","Data":"0ae1126331ace4b178fd1f6067fdd6c7db3a2db1dd916e12159936b16e62c868"} Dec 04 04:15:34 crc kubenswrapper[4806]: I1204 04:15:34.749145 4806 generic.go:334] "Generic (PLEG): container finished" podID="88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b" containerID="f5963e6d29ec3803ccbf9650a1ebe9d814b2db349acc32fccccdc1b1278f6bbe" exitCode=0 Dec 04 04:15:34 crc kubenswrapper[4806]: I1204 04:15:34.749548 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" event={"ID":"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b","Type":"ContainerDied","Data":"f5963e6d29ec3803ccbf9650a1ebe9d814b2db349acc32fccccdc1b1278f6bbe"} Dec 04 04:15:34 crc kubenswrapper[4806]: I1204 04:15:34.775152 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7496d77c9d-kjd5v" Dec 04 04:15:34 crc kubenswrapper[4806]: I1204 04:15:34.776386 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3d05a5b8-2294-4d30-a0b7-d865c8aae646","Type":"ContainerStarted","Data":"7ec503dfe6d0f406e74dc3d2e9f7a01dc76cebd7cd660728689128c515a833ae"} Dec 04 04:15:34 crc kubenswrapper[4806]: I1204 04:15:34.862251 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7496d77c9d-kjd5v"] Dec 04 04:15:34 crc kubenswrapper[4806]: I1204 04:15:34.867883 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7496d77c9d-kjd5v"] Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.075219 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.170834 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-config\") pod \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.170895 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4qg6\" (UniqueName: \"kubernetes.io/projected/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-kube-api-access-d4qg6\") pod \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.171066 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-dns-swift-storage-0\") pod \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.171105 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-dns-svc\") pod \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.171183 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-ovsdbserver-sb\") pod \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.171288 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-ovsdbserver-nb\") pod \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\" (UID: \"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b\") " Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.201991 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-kube-api-access-d4qg6" (OuterVolumeSpecName: "kube-api-access-d4qg6") pod "88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b" (UID: "88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b"). InnerVolumeSpecName "kube-api-access-d4qg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.274198 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4qg6\" (UniqueName: \"kubernetes.io/projected/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-kube-api-access-d4qg6\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.311892 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b" (UID: "88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.347833 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b" (UID: "88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.363553 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-config" (OuterVolumeSpecName: "config") pod "88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b" (UID: "88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.370003 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b" (UID: "88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.379044 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.379086 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.379097 4806 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.379106 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.384541 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b" (UID: "88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.437656 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6850510-af16-4d30-b30c-54a9bd2186d3" path="/var/lib/kubelet/pods/d6850510-af16-4d30-b30c-54a9bd2186d3/volumes" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.484786 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.599109 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5b8bd497fd-fhbpm" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.599461 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5b8bd497fd-fhbpm" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.800499 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a","Type":"ContainerStarted","Data":"7966f7061881708421f10d96b8080aa15dc08c8d80ce2f965e184afb8521d8f5"} Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.804772 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" event={"ID":"88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b","Type":"ContainerDied","Data":"cf6c16c3b16e4c90fc22041149c9f46639f91f7f32f8ff24db640101783a42eb"} Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.804853 4806 scope.go:117] "RemoveContainer" containerID="f5963e6d29ec3803ccbf9650a1ebe9d814b2db349acc32fccccdc1b1278f6bbe" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.805132 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-wbj9t" Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.855145 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3d05a5b8-2294-4d30-a0b7-d865c8aae646","Type":"ContainerStarted","Data":"1c8f67c0816b816c7aa1213b2bc51c67114270a0c9626945aab089d3dfd12fb8"} Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.858029 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-wbj9t"] Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.877448 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-wbj9t"] Dec 04 04:15:35 crc kubenswrapper[4806]: I1204 04:15:35.910813 4806 scope.go:117] "RemoveContainer" containerID="f9cd5f81df81499f791ff879391331a78bfc1dbd5a2e9964787d94f66ec4edfc" Dec 04 04:15:36 crc kubenswrapper[4806]: I1204 04:15:36.603174 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5b8bd497fd-fhbpm" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:36 crc kubenswrapper[4806]: I1204 04:15:36.603754 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5b8bd497fd-fhbpm" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:37 crc kubenswrapper[4806]: I1204 04:15:37.435377 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b" path="/var/lib/kubelet/pods/88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b/volumes" Dec 04 04:15:37 crc kubenswrapper[4806]: I1204 04:15:37.911045 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a","Type":"ContainerStarted","Data":"a025484d58c0a01fa8a7734430d72227df9a28eeb74d2849591d7d5d5d7d2a54"} Dec 04 04:15:37 crc kubenswrapper[4806]: I1204 04:15:37.912324 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 04:15:37 crc kubenswrapper[4806]: I1204 04:15:37.914121 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3d05a5b8-2294-4d30-a0b7-d865c8aae646","Type":"ContainerStarted","Data":"c135cbdabf5362fe9cd3307b6becccf833b9ff8729d075501cafae3adfe2a8f2"} Dec 04 04:15:37 crc kubenswrapper[4806]: I1204 04:15:37.914811 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 04 04:15:37 crc kubenswrapper[4806]: I1204 04:15:37.992527 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.992507177 podStartE2EDuration="5.992507177s" podCreationTimestamp="2025-12-04 04:15:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:15:37.983632916 +0000 UTC m=+1252.842145844" watchObservedRunningTime="2025-12-04 04:15:37.992507177 +0000 UTC m=+1252.851020105" Dec 04 04:15:37 crc kubenswrapper[4806]: I1204 04:15:37.993185 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.981500702 podStartE2EDuration="7.993175758s" podCreationTimestamp="2025-12-04 04:15:30 +0000 UTC" firstStartedPulling="2025-12-04 04:15:31.729074811 +0000 UTC m=+1246.587587739" lastFinishedPulling="2025-12-04 04:15:36.740749867 +0000 UTC m=+1251.599262795" observedRunningTime="2025-12-04 04:15:37.954770346 +0000 UTC m=+1252.813283274" watchObservedRunningTime="2025-12-04 04:15:37.993175758 +0000 UTC m=+1252.851688686" Dec 04 04:15:39 crc kubenswrapper[4806]: I1204 04:15:39.301914 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 04 04:15:39 crc kubenswrapper[4806]: I1204 04:15:39.371973 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 04:15:39 crc kubenswrapper[4806]: I1204 04:15:39.461527 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:39 crc kubenswrapper[4806]: I1204 04:15:39.698693 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-d9c876cd8-zfctt" Dec 04 04:15:39 crc kubenswrapper[4806]: I1204 04:15:39.937519 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" containerName="cinder-scheduler" containerID="cri-o://c7a92ed6ce652bbe78c097b6b9ee4658d9058abf96a0dcdded3a64674aa24633" gracePeriod=30 Dec 04 04:15:39 crc kubenswrapper[4806]: I1204 04:15:39.938147 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" containerName="probe" containerID="cri-o://c6109f8c356b8024f08644d074a1b5e27920bac999614a1ccac2d89260cc15d6" gracePeriod=30 Dec 04 04:15:40 crc kubenswrapper[4806]: I1204 04:15:40.669254 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5b8bd497fd-fhbpm" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:40 crc kubenswrapper[4806]: I1204 04:15:40.711248 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5b8bd497fd-fhbpm" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:40 crc kubenswrapper[4806]: I1204 04:15:40.948131 4806 generic.go:334] "Generic (PLEG): container finished" podID="83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" containerID="c6109f8c356b8024f08644d074a1b5e27920bac999614a1ccac2d89260cc15d6" exitCode=0 Dec 04 04:15:40 crc kubenswrapper[4806]: I1204 04:15:40.948173 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3","Type":"ContainerDied","Data":"c6109f8c356b8024f08644d074a1b5e27920bac999614a1ccac2d89260cc15d6"} Dec 04 04:15:41 crc kubenswrapper[4806]: I1204 04:15:41.644171 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5b8bd497fd-fhbpm" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:41 crc kubenswrapper[4806]: I1204 04:15:41.686122 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5b8bd497fd-fhbpm" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:41 crc kubenswrapper[4806]: I1204 04:15:41.699103 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:41 crc kubenswrapper[4806]: I1204 04:15:41.702633 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:42 crc kubenswrapper[4806]: I1204 04:15:42.470125 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-64cc7bf6f4-fcxhm" Dec 04 04:15:42 crc kubenswrapper[4806]: I1204 04:15:42.647097 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-949b6795d-z7dx2" podUID="ef61aa0b-2b2c-481f-b22d-4ea770a91711" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.163:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:42 crc kubenswrapper[4806]: I1204 04:15:42.654230 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-949b6795d-z7dx2" podUID="ef61aa0b-2b2c-481f-b22d-4ea770a91711" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.163:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:42 crc kubenswrapper[4806]: I1204 04:15:42.970778 4806 generic.go:334] "Generic (PLEG): container finished" podID="83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" containerID="c7a92ed6ce652bbe78c097b6b9ee4658d9058abf96a0dcdded3a64674aa24633" exitCode=0 Dec 04 04:15:42 crc kubenswrapper[4806]: I1204 04:15:42.971120 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3","Type":"ContainerDied","Data":"c7a92ed6ce652bbe78c097b6b9ee4658d9058abf96a0dcdded3a64674aa24633"} Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.479034 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.572529 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-etc-machine-id\") pod \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.572680 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-config-data-custom\") pod \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.572715 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-scripts\") pod \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.572771 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpfxl\" (UniqueName: \"kubernetes.io/projected/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-kube-api-access-lpfxl\") pod \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.572837 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-combined-ca-bundle\") pod \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.572880 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-config-data\") pod \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\" (UID: \"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3\") " Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.574254 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" (UID: "83c1ae2a-6cca-4ba1-8a30-e285d6023ba3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.590012 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-scripts" (OuterVolumeSpecName: "scripts") pod "83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" (UID: "83c1ae2a-6cca-4ba1-8a30-e285d6023ba3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.599201 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" (UID: "83c1ae2a-6cca-4ba1-8a30-e285d6023ba3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.601153 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-kube-api-access-lpfxl" (OuterVolumeSpecName: "kube-api-access-lpfxl") pod "83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" (UID: "83c1ae2a-6cca-4ba1-8a30-e285d6023ba3"). InnerVolumeSpecName "kube-api-access-lpfxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.651117 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-949b6795d-z7dx2" podUID="ef61aa0b-2b2c-481f-b22d-4ea770a91711" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.163:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.651475 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-949b6795d-z7dx2" podUID="ef61aa0b-2b2c-481f-b22d-4ea770a91711" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.163:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.676734 4806 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.676794 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.676806 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpfxl\" (UniqueName: \"kubernetes.io/projected/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-kube-api-access-lpfxl\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.676820 4806 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.746443 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" (UID: "83c1ae2a-6cca-4ba1-8a30-e285d6023ba3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.761388 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 04 04:15:43 crc kubenswrapper[4806]: E1204 04:15:43.761792 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6850510-af16-4d30-b30c-54a9bd2186d3" containerName="neutron-api" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.761812 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6850510-af16-4d30-b30c-54a9bd2186d3" containerName="neutron-api" Dec 04 04:15:43 crc kubenswrapper[4806]: E1204 04:15:43.761824 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b" containerName="init" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.761830 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b" containerName="init" Dec 04 04:15:43 crc kubenswrapper[4806]: E1204 04:15:43.761845 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" containerName="probe" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.761852 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" containerName="probe" Dec 04 04:15:43 crc kubenswrapper[4806]: E1204 04:15:43.761875 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" containerName="cinder-scheduler" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.761881 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" containerName="cinder-scheduler" Dec 04 04:15:43 crc kubenswrapper[4806]: E1204 04:15:43.761893 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b" containerName="dnsmasq-dns" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.761898 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b" containerName="dnsmasq-dns" Dec 04 04:15:43 crc kubenswrapper[4806]: E1204 04:15:43.761914 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6850510-af16-4d30-b30c-54a9bd2186d3" containerName="neutron-httpd" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.761937 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6850510-af16-4d30-b30c-54a9bd2186d3" containerName="neutron-httpd" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.762098 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" containerName="probe" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.762112 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="88c68b90-6bf2-4dfa-b17b-2ee9adc5be9b" containerName="dnsmasq-dns" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.762124 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6850510-af16-4d30-b30c-54a9bd2186d3" containerName="neutron-api" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.762134 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" containerName="cinder-scheduler" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.762149 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6850510-af16-4d30-b30c-54a9bd2186d3" containerName="neutron-httpd" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.762714 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.766442 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.769649 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-xgx8r" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.769793 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.779041 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.798679 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.884010 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8395825-5ec0-4682-b7e9-db184f570586-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e8395825-5ec0-4682-b7e9-db184f570586\") " pod="openstack/openstackclient" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.884138 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9txwk\" (UniqueName: \"kubernetes.io/projected/e8395825-5ec0-4682-b7e9-db184f570586-kube-api-access-9txwk\") pod \"openstackclient\" (UID: \"e8395825-5ec0-4682-b7e9-db184f570586\") " pod="openstack/openstackclient" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.884220 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e8395825-5ec0-4682-b7e9-db184f570586-openstack-config-secret\") pod \"openstackclient\" (UID: \"e8395825-5ec0-4682-b7e9-db184f570586\") " pod="openstack/openstackclient" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.884248 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e8395825-5ec0-4682-b7e9-db184f570586-openstack-config\") pod \"openstackclient\" (UID: \"e8395825-5ec0-4682-b7e9-db184f570586\") " pod="openstack/openstackclient" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.929078 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-config-data" (OuterVolumeSpecName: "config-data") pod "83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" (UID: "83c1ae2a-6cca-4ba1-8a30-e285d6023ba3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.989937 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e8395825-5ec0-4682-b7e9-db184f570586-openstack-config-secret\") pod \"openstackclient\" (UID: \"e8395825-5ec0-4682-b7e9-db184f570586\") " pod="openstack/openstackclient" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.989998 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e8395825-5ec0-4682-b7e9-db184f570586-openstack-config\") pod \"openstackclient\" (UID: \"e8395825-5ec0-4682-b7e9-db184f570586\") " pod="openstack/openstackclient" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.990029 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8395825-5ec0-4682-b7e9-db184f570586-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e8395825-5ec0-4682-b7e9-db184f570586\") " pod="openstack/openstackclient" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.990120 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9txwk\" (UniqueName: \"kubernetes.io/projected/e8395825-5ec0-4682-b7e9-db184f570586-kube-api-access-9txwk\") pod \"openstackclient\" (UID: \"e8395825-5ec0-4682-b7e9-db184f570586\") " pod="openstack/openstackclient" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.990211 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:43 crc kubenswrapper[4806]: I1204 04:15:43.993750 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e8395825-5ec0-4682-b7e9-db184f570586-openstack-config\") pod \"openstackclient\" (UID: \"e8395825-5ec0-4682-b7e9-db184f570586\") " pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.004697 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e8395825-5ec0-4682-b7e9-db184f570586-openstack-config-secret\") pod \"openstackclient\" (UID: \"e8395825-5ec0-4682-b7e9-db184f570586\") " pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.005839 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8395825-5ec0-4682-b7e9-db184f570586-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e8395825-5ec0-4682-b7e9-db184f570586\") " pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.022162 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"83c1ae2a-6cca-4ba1-8a30-e285d6023ba3","Type":"ContainerDied","Data":"0ef66db12f357783f39b9da9f04e16bb9344d54259932e83ac4010d2e56a21d9"} Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.022388 4806 scope.go:117] "RemoveContainer" containerID="c6109f8c356b8024f08644d074a1b5e27920bac999614a1ccac2d89260cc15d6" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.022606 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.024470 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9txwk\" (UniqueName: \"kubernetes.io/projected/e8395825-5ec0-4682-b7e9-db184f570586-kube-api-access-9txwk\") pod \"openstackclient\" (UID: \"e8395825-5ec0-4682-b7e9-db184f570586\") " pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.070660 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.118605 4806 scope.go:117] "RemoveContainer" containerID="c7a92ed6ce652bbe78c097b6b9ee4658d9058abf96a0dcdded3a64674aa24633" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.127140 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.255014 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.278999 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.293993 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.304990 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.306574 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.313201 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.313656 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.314815 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.321983 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.332302 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 04:15:44 crc kubenswrapper[4806]: E1204 04:15:44.435159 4806 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 04 04:15:44 crc kubenswrapper[4806]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_e8395825-5ec0-4682-b7e9-db184f570586_0(93bbab352da7be70fe956d2b8753955c25832a9654815213acfe42b56590c9b6): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"93bbab352da7be70fe956d2b8753955c25832a9654815213acfe42b56590c9b6" Netns:"/var/run/netns/37e7eaa2-c689-40ef-a846-6b73f8da4c8d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=93bbab352da7be70fe956d2b8753955c25832a9654815213acfe42b56590c9b6;K8S_POD_UID=e8395825-5ec0-4682-b7e9-db184f570586" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/e8395825-5ec0-4682-b7e9-db184f570586]: expected pod UID "e8395825-5ec0-4682-b7e9-db184f570586" but got "266dbc23-c65d-4b9d-978d-09f9de4a06e5" from Kube API Dec 04 04:15:44 crc kubenswrapper[4806]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 04 04:15:44 crc kubenswrapper[4806]: > Dec 04 04:15:44 crc kubenswrapper[4806]: E1204 04:15:44.435223 4806 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 04 04:15:44 crc kubenswrapper[4806]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_e8395825-5ec0-4682-b7e9-db184f570586_0(93bbab352da7be70fe956d2b8753955c25832a9654815213acfe42b56590c9b6): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"93bbab352da7be70fe956d2b8753955c25832a9654815213acfe42b56590c9b6" Netns:"/var/run/netns/37e7eaa2-c689-40ef-a846-6b73f8da4c8d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=93bbab352da7be70fe956d2b8753955c25832a9654815213acfe42b56590c9b6;K8S_POD_UID=e8395825-5ec0-4682-b7e9-db184f570586" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/e8395825-5ec0-4682-b7e9-db184f570586]: expected pod UID "e8395825-5ec0-4682-b7e9-db184f570586" but got "266dbc23-c65d-4b9d-978d-09f9de4a06e5" from Kube API Dec 04 04:15:44 crc kubenswrapper[4806]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 04 04:15:44 crc kubenswrapper[4806]: > pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.436168 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/266dbc23-c65d-4b9d-978d-09f9de4a06e5-openstack-config\") pod \"openstackclient\" (UID: \"266dbc23-c65d-4b9d-978d-09f9de4a06e5\") " pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.436299 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8205f150-c0db-445a-b005-30d0632610c0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.436424 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwxf7\" (UniqueName: \"kubernetes.io/projected/8205f150-c0db-445a-b005-30d0632610c0-kube-api-access-pwxf7\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.436549 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/266dbc23-c65d-4b9d-978d-09f9de4a06e5-openstack-config-secret\") pod \"openstackclient\" (UID: \"266dbc23-c65d-4b9d-978d-09f9de4a06e5\") " pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.436660 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8205f150-c0db-445a-b005-30d0632610c0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.436827 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8205f150-c0db-445a-b005-30d0632610c0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.436958 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8205f150-c0db-445a-b005-30d0632610c0-scripts\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.437100 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/266dbc23-c65d-4b9d-978d-09f9de4a06e5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"266dbc23-c65d-4b9d-978d-09f9de4a06e5\") " pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.437253 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84rkr\" (UniqueName: \"kubernetes.io/projected/266dbc23-c65d-4b9d-978d-09f9de4a06e5-kube-api-access-84rkr\") pod \"openstackclient\" (UID: \"266dbc23-c65d-4b9d-978d-09f9de4a06e5\") " pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.437365 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8205f150-c0db-445a-b005-30d0632610c0-config-data\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.539379 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/266dbc23-c65d-4b9d-978d-09f9de4a06e5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"266dbc23-c65d-4b9d-978d-09f9de4a06e5\") " pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.539701 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84rkr\" (UniqueName: \"kubernetes.io/projected/266dbc23-c65d-4b9d-978d-09f9de4a06e5-kube-api-access-84rkr\") pod \"openstackclient\" (UID: \"266dbc23-c65d-4b9d-978d-09f9de4a06e5\") " pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.539804 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8205f150-c0db-445a-b005-30d0632610c0-config-data\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.539974 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/266dbc23-c65d-4b9d-978d-09f9de4a06e5-openstack-config\") pod \"openstackclient\" (UID: \"266dbc23-c65d-4b9d-978d-09f9de4a06e5\") " pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.540059 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8205f150-c0db-445a-b005-30d0632610c0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.540145 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwxf7\" (UniqueName: \"kubernetes.io/projected/8205f150-c0db-445a-b005-30d0632610c0-kube-api-access-pwxf7\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.540223 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/266dbc23-c65d-4b9d-978d-09f9de4a06e5-openstack-config-secret\") pod \"openstackclient\" (UID: \"266dbc23-c65d-4b9d-978d-09f9de4a06e5\") " pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.540296 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8205f150-c0db-445a-b005-30d0632610c0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.540462 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8205f150-c0db-445a-b005-30d0632610c0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.540566 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8205f150-c0db-445a-b005-30d0632610c0-scripts\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.541561 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8205f150-c0db-445a-b005-30d0632610c0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.542099 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/266dbc23-c65d-4b9d-978d-09f9de4a06e5-openstack-config\") pod \"openstackclient\" (UID: \"266dbc23-c65d-4b9d-978d-09f9de4a06e5\") " pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.544455 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8205f150-c0db-445a-b005-30d0632610c0-scripts\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.545532 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8205f150-c0db-445a-b005-30d0632610c0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.546884 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8205f150-c0db-445a-b005-30d0632610c0-config-data\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.553442 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/266dbc23-c65d-4b9d-978d-09f9de4a06e5-openstack-config-secret\") pod \"openstackclient\" (UID: \"266dbc23-c65d-4b9d-978d-09f9de4a06e5\") " pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.554030 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/266dbc23-c65d-4b9d-978d-09f9de4a06e5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"266dbc23-c65d-4b9d-978d-09f9de4a06e5\") " pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.555579 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8205f150-c0db-445a-b005-30d0632610c0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.563692 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwxf7\" (UniqueName: \"kubernetes.io/projected/8205f150-c0db-445a-b005-30d0632610c0-kube-api-access-pwxf7\") pod \"cinder-scheduler-0\" (UID: \"8205f150-c0db-445a-b005-30d0632610c0\") " pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.568376 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84rkr\" (UniqueName: \"kubernetes.io/projected/266dbc23-c65d-4b9d-978d-09f9de4a06e5-kube-api-access-84rkr\") pod \"openstackclient\" (UID: \"266dbc23-c65d-4b9d-978d-09f9de4a06e5\") " pod="openstack/openstackclient" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.679220 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 04:15:44 crc kubenswrapper[4806]: I1204 04:15:44.688514 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.053016 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.066333 4806 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e8395825-5ec0-4682-b7e9-db184f570586" podUID="266dbc23-c65d-4b9d-978d-09f9de4a06e5" Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.081095 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.257950 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8395825-5ec0-4682-b7e9-db184f570586-combined-ca-bundle\") pod \"e8395825-5ec0-4682-b7e9-db184f570586\" (UID: \"e8395825-5ec0-4682-b7e9-db184f570586\") " Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.258064 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e8395825-5ec0-4682-b7e9-db184f570586-openstack-config\") pod \"e8395825-5ec0-4682-b7e9-db184f570586\" (UID: \"e8395825-5ec0-4682-b7e9-db184f570586\") " Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.258213 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e8395825-5ec0-4682-b7e9-db184f570586-openstack-config-secret\") pod \"e8395825-5ec0-4682-b7e9-db184f570586\" (UID: \"e8395825-5ec0-4682-b7e9-db184f570586\") " Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.258296 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9txwk\" (UniqueName: \"kubernetes.io/projected/e8395825-5ec0-4682-b7e9-db184f570586-kube-api-access-9txwk\") pod \"e8395825-5ec0-4682-b7e9-db184f570586\" (UID: \"e8395825-5ec0-4682-b7e9-db184f570586\") " Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.259485 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8395825-5ec0-4682-b7e9-db184f570586-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "e8395825-5ec0-4682-b7e9-db184f570586" (UID: "e8395825-5ec0-4682-b7e9-db184f570586"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.272260 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8395825-5ec0-4682-b7e9-db184f570586-kube-api-access-9txwk" (OuterVolumeSpecName: "kube-api-access-9txwk") pod "e8395825-5ec0-4682-b7e9-db184f570586" (UID: "e8395825-5ec0-4682-b7e9-db184f570586"). InnerVolumeSpecName "kube-api-access-9txwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.274101 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8395825-5ec0-4682-b7e9-db184f570586-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8395825-5ec0-4682-b7e9-db184f570586" (UID: "e8395825-5ec0-4682-b7e9-db184f570586"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.283497 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8395825-5ec0-4682-b7e9-db184f570586-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "e8395825-5ec0-4682-b7e9-db184f570586" (UID: "e8395825-5ec0-4682-b7e9-db184f570586"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.360562 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8395825-5ec0-4682-b7e9-db184f570586-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.360592 4806 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e8395825-5ec0-4682-b7e9-db184f570586-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.360601 4806 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e8395825-5ec0-4682-b7e9-db184f570586-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.360629 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9txwk\" (UniqueName: \"kubernetes.io/projected/e8395825-5ec0-4682-b7e9-db184f570586-kube-api-access-9txwk\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.421468 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.489003 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83c1ae2a-6cca-4ba1-8a30-e285d6023ba3" path="/var/lib/kubelet/pods/83c1ae2a-6cca-4ba1-8a30-e285d6023ba3/volumes" Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.489780 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8395825-5ec0-4682-b7e9-db184f570586" path="/var/lib/kubelet/pods/e8395825-5ec0-4682-b7e9-db184f570586/volumes" Dec 04 04:15:45 crc kubenswrapper[4806]: I1204 04:15:45.570894 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 04:15:46 crc kubenswrapper[4806]: I1204 04:15:46.092774 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"266dbc23-c65d-4b9d-978d-09f9de4a06e5","Type":"ContainerStarted","Data":"fc7798ca1af4ff439253d0bda2c661b7efa986617bc5b115df26284723ce4206"} Dec 04 04:15:46 crc kubenswrapper[4806]: I1204 04:15:46.094261 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 04:15:46 crc kubenswrapper[4806]: I1204 04:15:46.094841 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8205f150-c0db-445a-b005-30d0632610c0","Type":"ContainerStarted","Data":"34cdace5cb0915ede0d6879967a253c6482ffe3c7cef5db38573ba8d29340f3e"} Dec 04 04:15:46 crc kubenswrapper[4806]: I1204 04:15:46.106981 4806 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e8395825-5ec0-4682-b7e9-db184f570586" podUID="266dbc23-c65d-4b9d-978d-09f9de4a06e5" Dec 04 04:15:47 crc kubenswrapper[4806]: I1204 04:15:47.115310 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8205f150-c0db-445a-b005-30d0632610c0","Type":"ContainerStarted","Data":"30da34858176e66f52f0f362e52781f8b30fc7803ea3d0531a581c1e01e1a810"} Dec 04 04:15:47 crc kubenswrapper[4806]: I1204 04:15:47.153155 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="3d05a5b8-2294-4d30-a0b7-d865c8aae646" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.165:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:47 crc kubenswrapper[4806]: I1204 04:15:47.653146 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-949b6795d-z7dx2" podUID="ef61aa0b-2b2c-481f-b22d-4ea770a91711" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.163:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:47 crc kubenswrapper[4806]: I1204 04:15:47.660247 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-949b6795d-z7dx2" podUID="ef61aa0b-2b2c-481f-b22d-4ea770a91711" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.163:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:48 crc kubenswrapper[4806]: I1204 04:15:48.137579 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8205f150-c0db-445a-b005-30d0632610c0","Type":"ContainerStarted","Data":"2f81f15efd66082e460d8e618619780c9969b3f3cdf31c07993cbe42e9c7c0a7"} Dec 04 04:15:48 crc kubenswrapper[4806]: I1204 04:15:48.149163 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="3d05a5b8-2294-4d30-a0b7-d865c8aae646" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.165:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:48 crc kubenswrapper[4806]: I1204 04:15:48.181417 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.181399084 podStartE2EDuration="4.181399084s" podCreationTimestamp="2025-12-04 04:15:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:15:48.170859511 +0000 UTC m=+1263.029372449" watchObservedRunningTime="2025-12-04 04:15:48.181399084 +0000 UTC m=+1263.039912012" Dec 04 04:15:48 crc kubenswrapper[4806]: I1204 04:15:48.698144 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-949b6795d-z7dx2" podUID="ef61aa0b-2b2c-481f-b22d-4ea770a91711" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.163:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:48 crc kubenswrapper[4806]: I1204 04:15:48.698156 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-949b6795d-z7dx2" podUID="ef61aa0b-2b2c-481f-b22d-4ea770a91711" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.163:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:48 crc kubenswrapper[4806]: I1204 04:15:48.726379 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:48 crc kubenswrapper[4806]: I1204 04:15:48.761096 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-949b6795d-z7dx2" Dec 04 04:15:48 crc kubenswrapper[4806]: I1204 04:15:48.827202 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5b8bd497fd-fhbpm"] Dec 04 04:15:48 crc kubenswrapper[4806]: I1204 04:15:48.827621 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5b8bd497fd-fhbpm" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api-log" containerID="cri-o://b84ee829c765879ba892935281145d03c4ca17da2bcde8164c8b78105e71d72e" gracePeriod=30 Dec 04 04:15:48 crc kubenswrapper[4806]: I1204 04:15:48.827857 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5b8bd497fd-fhbpm" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api" containerID="cri-o://44c099c1b1e2667cc7a3c3a248b3234946f9dc0e882330b14185ec61de016538" gracePeriod=30 Dec 04 04:15:49 crc kubenswrapper[4806]: I1204 04:15:49.151656 4806 generic.go:334] "Generic (PLEG): container finished" podID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerID="b84ee829c765879ba892935281145d03c4ca17da2bcde8164c8b78105e71d72e" exitCode=143 Dec 04 04:15:49 crc kubenswrapper[4806]: I1204 04:15:49.151767 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b8bd497fd-fhbpm" event={"ID":"1c395cfa-4cad-40eb-8362-cb2819fa2024","Type":"ContainerDied","Data":"b84ee829c765879ba892935281145d03c4ca17da2bcde8164c8b78105e71d72e"} Dec 04 04:15:49 crc kubenswrapper[4806]: I1204 04:15:49.679596 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 04 04:15:52 crc kubenswrapper[4806]: I1204 04:15:52.333248 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5b8bd497fd-fhbpm" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:59810->10.217.0.159:9311: read: connection reset by peer" Dec 04 04:15:52 crc kubenswrapper[4806]: I1204 04:15:52.333305 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5b8bd497fd-fhbpm" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:59822->10.217.0.159:9311: read: connection reset by peer" Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.153188 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="3d05a5b8-2294-4d30-a0b7-d865c8aae646" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.165:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.210764 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.221192 4806 generic.go:334] "Generic (PLEG): container finished" podID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerID="44c099c1b1e2667cc7a3c3a248b3234946f9dc0e882330b14185ec61de016538" exitCode=0 Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.221231 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b8bd497fd-fhbpm" event={"ID":"1c395cfa-4cad-40eb-8362-cb2819fa2024","Type":"ContainerDied","Data":"44c099c1b1e2667cc7a3c3a248b3234946f9dc0e882330b14185ec61de016538"} Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.221296 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5b8bd497fd-fhbpm" event={"ID":"1c395cfa-4cad-40eb-8362-cb2819fa2024","Type":"ContainerDied","Data":"63648aa25b95162a262f332f2c4fb8d34c5271959c7089456a50dfb0e8e5bf64"} Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.221311 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63648aa25b95162a262f332f2c4fb8d34c5271959c7089456a50dfb0e8e5bf64" Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.271151 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.368627 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9l84g\" (UniqueName: \"kubernetes.io/projected/1c395cfa-4cad-40eb-8362-cb2819fa2024-kube-api-access-9l84g\") pod \"1c395cfa-4cad-40eb-8362-cb2819fa2024\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.368990 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-config-data-custom\") pod \"1c395cfa-4cad-40eb-8362-cb2819fa2024\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.369020 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c395cfa-4cad-40eb-8362-cb2819fa2024-logs\") pod \"1c395cfa-4cad-40eb-8362-cb2819fa2024\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.369154 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-config-data\") pod \"1c395cfa-4cad-40eb-8362-cb2819fa2024\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.369221 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-combined-ca-bundle\") pod \"1c395cfa-4cad-40eb-8362-cb2819fa2024\" (UID: \"1c395cfa-4cad-40eb-8362-cb2819fa2024\") " Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.370352 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c395cfa-4cad-40eb-8362-cb2819fa2024-logs" (OuterVolumeSpecName: "logs") pod "1c395cfa-4cad-40eb-8362-cb2819fa2024" (UID: "1c395cfa-4cad-40eb-8362-cb2819fa2024"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.386297 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c395cfa-4cad-40eb-8362-cb2819fa2024-kube-api-access-9l84g" (OuterVolumeSpecName: "kube-api-access-9l84g") pod "1c395cfa-4cad-40eb-8362-cb2819fa2024" (UID: "1c395cfa-4cad-40eb-8362-cb2819fa2024"). InnerVolumeSpecName "kube-api-access-9l84g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.386658 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1c395cfa-4cad-40eb-8362-cb2819fa2024" (UID: "1c395cfa-4cad-40eb-8362-cb2819fa2024"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.450108 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1c395cfa-4cad-40eb-8362-cb2819fa2024" (UID: "1c395cfa-4cad-40eb-8362-cb2819fa2024"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.471752 4806 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.471786 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c395cfa-4cad-40eb-8362-cb2819fa2024-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.471796 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.471806 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9l84g\" (UniqueName: \"kubernetes.io/projected/1c395cfa-4cad-40eb-8362-cb2819fa2024-kube-api-access-9l84g\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.504049 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-config-data" (OuterVolumeSpecName: "config-data") pod "1c395cfa-4cad-40eb-8362-cb2819fa2024" (UID: "1c395cfa-4cad-40eb-8362-cb2819fa2024"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:15:53 crc kubenswrapper[4806]: I1204 04:15:53.574260 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c395cfa-4cad-40eb-8362-cb2819fa2024-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:15:54 crc kubenswrapper[4806]: I1204 04:15:54.232761 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5b8bd497fd-fhbpm" Dec 04 04:15:54 crc kubenswrapper[4806]: I1204 04:15:54.286068 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5b8bd497fd-fhbpm"] Dec 04 04:15:54 crc kubenswrapper[4806]: I1204 04:15:54.301950 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5b8bd497fd-fhbpm"] Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.066274 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.366991 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-685c486f6f-6zmmj"] Dec 04 04:15:55 crc kubenswrapper[4806]: E1204 04:15:55.367420 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api-log" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.367441 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api-log" Dec 04 04:15:55 crc kubenswrapper[4806]: E1204 04:15:55.367455 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.367462 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.367684 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.367713 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" containerName="barbican-api-log" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.368742 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.373222 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.376366 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.377179 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.388654 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-685c486f6f-6zmmj"] Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.428718 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/794d2127-4482-4196-b558-d24f003e96ca-etc-swift\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.429108 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x859b\" (UniqueName: \"kubernetes.io/projected/794d2127-4482-4196-b558-d24f003e96ca-kube-api-access-x859b\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.429224 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/794d2127-4482-4196-b558-d24f003e96ca-run-httpd\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.429321 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/794d2127-4482-4196-b558-d24f003e96ca-public-tls-certs\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.429390 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/794d2127-4482-4196-b558-d24f003e96ca-combined-ca-bundle\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.429550 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/794d2127-4482-4196-b558-d24f003e96ca-log-httpd\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.429728 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/794d2127-4482-4196-b558-d24f003e96ca-internal-tls-certs\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.430975 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/794d2127-4482-4196-b558-d24f003e96ca-config-data\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.442784 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c395cfa-4cad-40eb-8362-cb2819fa2024" path="/var/lib/kubelet/pods/1c395cfa-4cad-40eb-8362-cb2819fa2024/volumes" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.532265 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/794d2127-4482-4196-b558-d24f003e96ca-log-httpd\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.532342 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/794d2127-4482-4196-b558-d24f003e96ca-internal-tls-certs\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.532410 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/794d2127-4482-4196-b558-d24f003e96ca-config-data\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.532455 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/794d2127-4482-4196-b558-d24f003e96ca-etc-swift\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.532481 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x859b\" (UniqueName: \"kubernetes.io/projected/794d2127-4482-4196-b558-d24f003e96ca-kube-api-access-x859b\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.532520 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/794d2127-4482-4196-b558-d24f003e96ca-run-httpd\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.532560 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/794d2127-4482-4196-b558-d24f003e96ca-public-tls-certs\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.532589 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/794d2127-4482-4196-b558-d24f003e96ca-combined-ca-bundle\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.532784 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/794d2127-4482-4196-b558-d24f003e96ca-log-httpd\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.533427 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/794d2127-4482-4196-b558-d24f003e96ca-run-httpd\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.557550 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/794d2127-4482-4196-b558-d24f003e96ca-internal-tls-certs\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.557864 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/794d2127-4482-4196-b558-d24f003e96ca-config-data\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.559456 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/794d2127-4482-4196-b558-d24f003e96ca-public-tls-certs\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.560220 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/794d2127-4482-4196-b558-d24f003e96ca-etc-swift\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.562768 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/794d2127-4482-4196-b558-d24f003e96ca-combined-ca-bundle\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.572002 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x859b\" (UniqueName: \"kubernetes.io/projected/794d2127-4482-4196-b558-d24f003e96ca-kube-api-access-x859b\") pod \"swift-proxy-685c486f6f-6zmmj\" (UID: \"794d2127-4482-4196-b558-d24f003e96ca\") " pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:55 crc kubenswrapper[4806]: I1204 04:15:55.687167 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:15:57 crc kubenswrapper[4806]: I1204 04:15:57.046783 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:15:57 crc kubenswrapper[4806]: I1204 04:15:57.046837 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:15:58 crc kubenswrapper[4806]: I1204 04:15:58.783193 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:15:58 crc kubenswrapper[4806]: I1204 04:15:58.790191 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="ceilometer-central-agent" containerID="cri-o://60c3830bf09243bd142f2e2e096840ef7ccc4dc359af29a53725a6282a2334e7" gracePeriod=30 Dec 04 04:15:58 crc kubenswrapper[4806]: I1204 04:15:58.791331 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="ceilometer-notification-agent" containerID="cri-o://0ae1126331ace4b178fd1f6067fdd6c7db3a2db1dd916e12159936b16e62c868" gracePeriod=30 Dec 04 04:15:58 crc kubenswrapper[4806]: I1204 04:15:58.791332 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="sg-core" containerID="cri-o://7966f7061881708421f10d96b8080aa15dc08c8d80ce2f965e184afb8521d8f5" gracePeriod=30 Dec 04 04:15:58 crc kubenswrapper[4806]: I1204 04:15:58.791529 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="proxy-httpd" containerID="cri-o://a025484d58c0a01fa8a7734430d72227df9a28eeb74d2849591d7d5d5d7d2a54" gracePeriod=30 Dec 04 04:15:58 crc kubenswrapper[4806]: I1204 04:15:58.809768 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.164:3000/\": EOF" Dec 04 04:15:59 crc kubenswrapper[4806]: I1204 04:15:59.314861 4806 generic.go:334] "Generic (PLEG): container finished" podID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerID="a025484d58c0a01fa8a7734430d72227df9a28eeb74d2849591d7d5d5d7d2a54" exitCode=0 Dec 04 04:15:59 crc kubenswrapper[4806]: I1204 04:15:59.315294 4806 generic.go:334] "Generic (PLEG): container finished" podID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerID="7966f7061881708421f10d96b8080aa15dc08c8d80ce2f965e184afb8521d8f5" exitCode=2 Dec 04 04:15:59 crc kubenswrapper[4806]: I1204 04:15:59.315308 4806 generic.go:334] "Generic (PLEG): container finished" podID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerID="60c3830bf09243bd142f2e2e096840ef7ccc4dc359af29a53725a6282a2334e7" exitCode=0 Dec 04 04:15:59 crc kubenswrapper[4806]: I1204 04:15:59.314954 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a","Type":"ContainerDied","Data":"a025484d58c0a01fa8a7734430d72227df9a28eeb74d2849591d7d5d5d7d2a54"} Dec 04 04:15:59 crc kubenswrapper[4806]: I1204 04:15:59.315369 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a","Type":"ContainerDied","Data":"7966f7061881708421f10d96b8080aa15dc08c8d80ce2f965e184afb8521d8f5"} Dec 04 04:15:59 crc kubenswrapper[4806]: I1204 04:15:59.315390 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a","Type":"ContainerDied","Data":"60c3830bf09243bd142f2e2e096840ef7ccc4dc359af29a53725a6282a2334e7"} Dec 04 04:16:00 crc kubenswrapper[4806]: I1204 04:16:00.326546 4806 generic.go:334] "Generic (PLEG): container finished" podID="2265394b-be59-4797-b54d-369e5a5624e0" containerID="33669b61f9a2e337d989305386e72fc8b569e4ce6dbdb13e37c0aacc3aaf1ca1" exitCode=137 Dec 04 04:16:00 crc kubenswrapper[4806]: I1204 04:16:00.326701 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc4f74d8d-xj2xn" event={"ID":"2265394b-be59-4797-b54d-369e5a5624e0","Type":"ContainerDied","Data":"33669b61f9a2e337d989305386e72fc8b569e4ce6dbdb13e37c0aacc3aaf1ca1"} Dec 04 04:16:00 crc kubenswrapper[4806]: I1204 04:16:00.332837 4806 generic.go:334] "Generic (PLEG): container finished" podID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerID="0ae1126331ace4b178fd1f6067fdd6c7db3a2db1dd916e12159936b16e62c868" exitCode=0 Dec 04 04:16:00 crc kubenswrapper[4806]: I1204 04:16:00.332875 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a","Type":"ContainerDied","Data":"0ae1126331ace4b178fd1f6067fdd6c7db3a2db1dd916e12159936b16e62c868"} Dec 04 04:16:00 crc kubenswrapper[4806]: I1204 04:16:00.840918 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.164:3000/\": dial tcp 10.217.0.164:3000: connect: connection refused" Dec 04 04:16:01 crc kubenswrapper[4806]: I1204 04:16:01.342368 4806 generic.go:334] "Generic (PLEG): container finished" podID="d2d53864-c3c8-4e40-8791-0bac49ba561e" containerID="a440aa262b334c36c507fc4d0fece95b286074ca6be74690b931e0ac6569d085" exitCode=137 Dec 04 04:16:01 crc kubenswrapper[4806]: I1204 04:16:01.342719 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d8bd6c96d-4sqhf" event={"ID":"d2d53864-c3c8-4e40-8791-0bac49ba561e","Type":"ContainerDied","Data":"a440aa262b334c36c507fc4d0fece95b286074ca6be74690b931e0ac6569d085"} Dec 04 04:16:03 crc kubenswrapper[4806]: I1204 04:16:03.712522 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 04:16:03 crc kubenswrapper[4806]: I1204 04:16:03.712820 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="06eb71d9-0ccb-4cee-8217-97dae82cbcdf" containerName="glance-log" containerID="cri-o://39f515cae2496194a873198e9bcc3ae7bc4668ff452c27135e6987ebf4276afc" gracePeriod=30 Dec 04 04:16:03 crc kubenswrapper[4806]: I1204 04:16:03.713379 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="06eb71d9-0ccb-4cee-8217-97dae82cbcdf" containerName="glance-httpd" containerID="cri-o://8428b43f132768c89eff84a87a2ffb8f9182b7bbc7b553fed98c07bff6da1ba9" gracePeriod=30 Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.130233 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.309740 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-run-httpd\") pod \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.309835 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-config-data\") pod \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.309861 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-log-httpd\") pod \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.309899 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd7kr\" (UniqueName: \"kubernetes.io/projected/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-kube-api-access-cd7kr\") pod \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.309963 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-combined-ca-bundle\") pod \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.309984 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-sg-core-conf-yaml\") pod \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.310060 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-scripts\") pod \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\" (UID: \"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a\") " Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.311239 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" (UID: "b7003c3f-33df-4ef5-b58f-83aa0b80cf3a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.313120 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" (UID: "b7003c3f-33df-4ef5-b58f-83aa0b80cf3a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.317051 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-kube-api-access-cd7kr" (OuterVolumeSpecName: "kube-api-access-cd7kr") pod "b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" (UID: "b7003c3f-33df-4ef5-b58f-83aa0b80cf3a"). InnerVolumeSpecName "kube-api-access-cd7kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.317187 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-scripts" (OuterVolumeSpecName: "scripts") pod "b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" (UID: "b7003c3f-33df-4ef5-b58f-83aa0b80cf3a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.344056 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" (UID: "b7003c3f-33df-4ef5-b58f-83aa0b80cf3a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.367985 4806 generic.go:334] "Generic (PLEG): container finished" podID="06eb71d9-0ccb-4cee-8217-97dae82cbcdf" containerID="39f515cae2496194a873198e9bcc3ae7bc4668ff452c27135e6987ebf4276afc" exitCode=143 Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.368041 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"06eb71d9-0ccb-4cee-8217-97dae82cbcdf","Type":"ContainerDied","Data":"39f515cae2496194a873198e9bcc3ae7bc4668ff452c27135e6987ebf4276afc"} Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.387422 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d8bd6c96d-4sqhf" event={"ID":"d2d53864-c3c8-4e40-8791-0bac49ba561e","Type":"ContainerStarted","Data":"3d31a9a0fb8d228e6b10c413871aec203543909646c4d72d75fbdde129623505"} Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.400945 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"266dbc23-c65d-4b9d-978d-09f9de4a06e5","Type":"ContainerStarted","Data":"baaff4511a74b6e7716408488cf8fd6dba86bd59aa0d3d541cfda942deee85b0"} Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.412303 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc4f74d8d-xj2xn" event={"ID":"2265394b-be59-4797-b54d-369e5a5624e0","Type":"ContainerStarted","Data":"40e0a10f18db6fa40112391f3276df8272d8be2efe99e6d7b0a5ae22be825bb3"} Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.415496 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b7003c3f-33df-4ef5-b58f-83aa0b80cf3a","Type":"ContainerDied","Data":"b613d6faf3ea89efa3ad934817382dc7be856928ea65fdbe14fcc18301077440"} Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.416178 4806 scope.go:117] "RemoveContainer" containerID="a025484d58c0a01fa8a7734430d72227df9a28eeb74d2849591d7d5d5d7d2a54" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.415754 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.418780 4806 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.419046 4806 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.419127 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd7kr\" (UniqueName: \"kubernetes.io/projected/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-kube-api-access-cd7kr\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.419196 4806 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.419260 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.464116 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" (UID: "b7003c3f-33df-4ef5-b58f-83aa0b80cf3a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.497874 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.347485223 podStartE2EDuration="20.4978485s" podCreationTimestamp="2025-12-04 04:15:44 +0000 UTC" firstStartedPulling="2025-12-04 04:15:45.446175652 +0000 UTC m=+1260.304688590" lastFinishedPulling="2025-12-04 04:16:03.596538939 +0000 UTC m=+1278.455051867" observedRunningTime="2025-12-04 04:16:04.450432314 +0000 UTC m=+1279.308945232" watchObservedRunningTime="2025-12-04 04:16:04.4978485 +0000 UTC m=+1279.356361448" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.524312 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.534432 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-685c486f6f-6zmmj"] Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.571048 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-config-data" (OuterVolumeSpecName: "config-data") pod "b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" (UID: "b7003c3f-33df-4ef5-b58f-83aa0b80cf3a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.625881 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.656347 4806 scope.go:117] "RemoveContainer" containerID="7966f7061881708421f10d96b8080aa15dc08c8d80ce2f965e184afb8521d8f5" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.689219 4806 scope.go:117] "RemoveContainer" containerID="0ae1126331ace4b178fd1f6067fdd6c7db3a2db1dd916e12159936b16e62c868" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.713987 4806 scope.go:117] "RemoveContainer" containerID="60c3830bf09243bd142f2e2e096840ef7ccc4dc359af29a53725a6282a2334e7" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.774426 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.789946 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.805232 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:16:04 crc kubenswrapper[4806]: E1204 04:16:04.805873 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="ceilometer-central-agent" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.806004 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="ceilometer-central-agent" Dec 04 04:16:04 crc kubenswrapper[4806]: E1204 04:16:04.806126 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="sg-core" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.806203 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="sg-core" Dec 04 04:16:04 crc kubenswrapper[4806]: E1204 04:16:04.806287 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="proxy-httpd" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.806361 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="proxy-httpd" Dec 04 04:16:04 crc kubenswrapper[4806]: E1204 04:16:04.806440 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="ceilometer-notification-agent" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.806525 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="ceilometer-notification-agent" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.806867 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="ceilometer-central-agent" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.806984 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="ceilometer-notification-agent" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.807083 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="sg-core" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.807181 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" containerName="proxy-httpd" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.809393 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.811712 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.824722 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.859352 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.961350 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/482c5aef-993c-44ea-af1b-60e757b344a1-log-httpd\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.961396 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.961447 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-config-data\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.961488 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/482c5aef-993c-44ea-af1b-60e757b344a1-run-httpd\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.961536 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd2qw\" (UniqueName: \"kubernetes.io/projected/482c5aef-993c-44ea-af1b-60e757b344a1-kube-api-access-sd2qw\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.961581 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-scripts\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:04 crc kubenswrapper[4806]: I1204 04:16:04.961606 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.063082 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-scripts\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.063314 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.063408 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/482c5aef-993c-44ea-af1b-60e757b344a1-log-httpd\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.063513 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.063614 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-config-data\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.063707 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/482c5aef-993c-44ea-af1b-60e757b344a1-run-httpd\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.063808 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd2qw\" (UniqueName: \"kubernetes.io/projected/482c5aef-993c-44ea-af1b-60e757b344a1-kube-api-access-sd2qw\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.065161 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/482c5aef-993c-44ea-af1b-60e757b344a1-run-httpd\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.065210 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/482c5aef-993c-44ea-af1b-60e757b344a1-log-httpd\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.069740 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-scripts\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.070560 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-config-data\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.072360 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.074692 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.087884 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd2qw\" (UniqueName: \"kubernetes.io/projected/482c5aef-993c-44ea-af1b-60e757b344a1-kube-api-access-sd2qw\") pod \"ceilometer-0\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " pod="openstack/ceilometer-0" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.180254 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.446776 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7003c3f-33df-4ef5-b58f-83aa0b80cf3a" path="/var/lib/kubelet/pods/b7003c3f-33df-4ef5-b58f-83aa0b80cf3a/volumes" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.454331 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.454370 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.454383 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-685c486f6f-6zmmj" event={"ID":"794d2127-4482-4196-b558-d24f003e96ca","Type":"ContainerStarted","Data":"d266acd3b7b17f28b02936915ba57cbfc707f01558ed2d6f3770c5047d267f99"} Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.454398 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-685c486f6f-6zmmj" event={"ID":"794d2127-4482-4196-b558-d24f003e96ca","Type":"ContainerStarted","Data":"8b2191accf73c940550f14694fd293ca2537dfeb83cca4e0805f1d01eec382cb"} Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.454409 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-685c486f6f-6zmmj" event={"ID":"794d2127-4482-4196-b558-d24f003e96ca","Type":"ContainerStarted","Data":"20165437bcea16c909f963b340d5954d599a73a0c5635fe5501a171f9ebe9431"} Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.504697 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-685c486f6f-6zmmj" podStartSLOduration=10.504674041 podStartE2EDuration="10.504674041s" podCreationTimestamp="2025-12-04 04:15:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:16:05.500487998 +0000 UTC m=+1280.359000916" watchObservedRunningTime="2025-12-04 04:16:05.504674041 +0000 UTC m=+1280.363186969" Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.520764 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.751123 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.909653 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.909983 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="60587700-0430-4cc5-a572-234a2a946748" containerName="glance-log" containerID="cri-o://0ab275c492bba0348963efaa7a72714cbf7dc07e88ae48f2852a33149e3aa1b2" gracePeriod=30 Dec 04 04:16:05 crc kubenswrapper[4806]: I1204 04:16:05.910067 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="60587700-0430-4cc5-a572-234a2a946748" containerName="glance-httpd" containerID="cri-o://995f3e6bceddd25455c101aa2af9f6e81db185f21e393da766f60b43f9a63f19" gracePeriod=30 Dec 04 04:16:06 crc kubenswrapper[4806]: I1204 04:16:06.451463 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"482c5aef-993c-44ea-af1b-60e757b344a1","Type":"ContainerStarted","Data":"63494d71f85fc8f5e64a02a33ddba2f31064442d8463ae1a86596e0e2c895b7d"} Dec 04 04:16:06 crc kubenswrapper[4806]: I1204 04:16:06.453896 4806 generic.go:334] "Generic (PLEG): container finished" podID="60587700-0430-4cc5-a572-234a2a946748" containerID="0ab275c492bba0348963efaa7a72714cbf7dc07e88ae48f2852a33149e3aa1b2" exitCode=143 Dec 04 04:16:06 crc kubenswrapper[4806]: I1204 04:16:06.454614 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"60587700-0430-4cc5-a572-234a2a946748","Type":"ContainerDied","Data":"0ab275c492bba0348963efaa7a72714cbf7dc07e88ae48f2852a33149e3aa1b2"} Dec 04 04:16:08 crc kubenswrapper[4806]: I1204 04:16:08.174847 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="06eb71d9-0ccb-4cee-8217-97dae82cbcdf" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.154:9292/healthcheck\": read tcp 10.217.0.2:39440->10.217.0.154:9292: read: connection reset by peer" Dec 04 04:16:08 crc kubenswrapper[4806]: I1204 04:16:08.175155 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="06eb71d9-0ccb-4cee-8217-97dae82cbcdf" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.154:9292/healthcheck\": read tcp 10.217.0.2:39426->10.217.0.154:9292: read: connection reset by peer" Dec 04 04:16:08 crc kubenswrapper[4806]: I1204 04:16:08.473302 4806 generic.go:334] "Generic (PLEG): container finished" podID="06eb71d9-0ccb-4cee-8217-97dae82cbcdf" containerID="8428b43f132768c89eff84a87a2ffb8f9182b7bbc7b553fed98c07bff6da1ba9" exitCode=0 Dec 04 04:16:08 crc kubenswrapper[4806]: I1204 04:16:08.473391 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"06eb71d9-0ccb-4cee-8217-97dae82cbcdf","Type":"ContainerDied","Data":"8428b43f132768c89eff84a87a2ffb8f9182b7bbc7b553fed98c07bff6da1ba9"} Dec 04 04:16:08 crc kubenswrapper[4806]: I1204 04:16:08.475880 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"482c5aef-993c-44ea-af1b-60e757b344a1","Type":"ContainerStarted","Data":"7ad61da1634012fe21b7a95c9f2df9d3c3c0d4d7ccff7962a606adf6f855e7c3"} Dec 04 04:16:08 crc kubenswrapper[4806]: I1204 04:16:08.475950 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"482c5aef-993c-44ea-af1b-60e757b344a1","Type":"ContainerStarted","Data":"773b26bf87027564a689deaa13d0650d55c06518588cdf0ab050f2ab25fe4475"} Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.508232 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"06eb71d9-0ccb-4cee-8217-97dae82cbcdf","Type":"ContainerDied","Data":"8bdbdd7a083cb0978851ed58cc23dffb52dd14e942e37541a37efa80260ebbff"} Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.508669 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bdbdd7a083cb0978851ed58cc23dffb52dd14e942e37541a37efa80260ebbff" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.517197 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"482c5aef-993c-44ea-af1b-60e757b344a1","Type":"ContainerStarted","Data":"2e69d9c7597e4ceef4adc4d6f023fe5bbc10eca0f383bad48be756d435168c3f"} Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.563381 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.685796 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-config-data\") pod \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.685916 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-scripts\") pod \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.685970 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-logs\") pod \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.686039 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-internal-tls-certs\") pod \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.686535 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-logs" (OuterVolumeSpecName: "logs") pod "06eb71d9-0ccb-4cee-8217-97dae82cbcdf" (UID: "06eb71d9-0ccb-4cee-8217-97dae82cbcdf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.686583 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-httpd-run\") pod \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.686617 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-combined-ca-bundle\") pod \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.686654 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.686689 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfwnk\" (UniqueName: \"kubernetes.io/projected/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-kube-api-access-lfwnk\") pod \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.687058 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.688153 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "06eb71d9-0ccb-4cee-8217-97dae82cbcdf" (UID: "06eb71d9-0ccb-4cee-8217-97dae82cbcdf"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.708148 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-kube-api-access-lfwnk" (OuterVolumeSpecName: "kube-api-access-lfwnk") pod "06eb71d9-0ccb-4cee-8217-97dae82cbcdf" (UID: "06eb71d9-0ccb-4cee-8217-97dae82cbcdf"). InnerVolumeSpecName "kube-api-access-lfwnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.712705 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "06eb71d9-0ccb-4cee-8217-97dae82cbcdf" (UID: "06eb71d9-0ccb-4cee-8217-97dae82cbcdf"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.722597 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-scripts" (OuterVolumeSpecName: "scripts") pod "06eb71d9-0ccb-4cee-8217-97dae82cbcdf" (UID: "06eb71d9-0ccb-4cee-8217-97dae82cbcdf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.787338 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06eb71d9-0ccb-4cee-8217-97dae82cbcdf" (UID: "06eb71d9-0ccb-4cee-8217-97dae82cbcdf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.788304 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-combined-ca-bundle\") pod \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.788881 4806 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.788970 4806 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.789027 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfwnk\" (UniqueName: \"kubernetes.io/projected/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-kube-api-access-lfwnk\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.789081 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:09 crc kubenswrapper[4806]: W1204 04:16:09.791011 4806 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/06eb71d9-0ccb-4cee-8217-97dae82cbcdf/volumes/kubernetes.io~secret/combined-ca-bundle Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.791048 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06eb71d9-0ccb-4cee-8217-97dae82cbcdf" (UID: "06eb71d9-0ccb-4cee-8217-97dae82cbcdf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.834033 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "06eb71d9-0ccb-4cee-8217-97dae82cbcdf" (UID: "06eb71d9-0ccb-4cee-8217-97dae82cbcdf"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.849801 4806 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.895290 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-config-data" (OuterVolumeSpecName: "config-data") pod "06eb71d9-0ccb-4cee-8217-97dae82cbcdf" (UID: "06eb71d9-0ccb-4cee-8217-97dae82cbcdf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.895668 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-config-data\") pod \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\" (UID: \"06eb71d9-0ccb-4cee-8217-97dae82cbcdf\") " Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.896250 4806 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.896268 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.896276 4806 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:09 crc kubenswrapper[4806]: W1204 04:16:09.896352 4806 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/06eb71d9-0ccb-4cee-8217-97dae82cbcdf/volumes/kubernetes.io~secret/config-data Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.896363 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-config-data" (OuterVolumeSpecName: "config-data") pod "06eb71d9-0ccb-4cee-8217-97dae82cbcdf" (UID: "06eb71d9-0ccb-4cee-8217-97dae82cbcdf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.951044 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.952229 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:16:09 crc kubenswrapper[4806]: I1204 04:16:09.997517 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06eb71d9-0ccb-4cee-8217-97dae82cbcdf-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.138029 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.205470 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-scripts\") pod \"60587700-0430-4cc5-a572-234a2a946748\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.205584 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-combined-ca-bundle\") pod \"60587700-0430-4cc5-a572-234a2a946748\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.205634 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"60587700-0430-4cc5-a572-234a2a946748\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.205677 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/60587700-0430-4cc5-a572-234a2a946748-httpd-run\") pod \"60587700-0430-4cc5-a572-234a2a946748\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.205703 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-public-tls-certs\") pod \"60587700-0430-4cc5-a572-234a2a946748\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.205721 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60587700-0430-4cc5-a572-234a2a946748-logs\") pod \"60587700-0430-4cc5-a572-234a2a946748\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.205816 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-config-data\") pod \"60587700-0430-4cc5-a572-234a2a946748\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.205836 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxkjq\" (UniqueName: \"kubernetes.io/projected/60587700-0430-4cc5-a572-234a2a946748-kube-api-access-qxkjq\") pod \"60587700-0430-4cc5-a572-234a2a946748\" (UID: \"60587700-0430-4cc5-a572-234a2a946748\") " Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.210360 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60587700-0430-4cc5-a572-234a2a946748-kube-api-access-qxkjq" (OuterVolumeSpecName: "kube-api-access-qxkjq") pod "60587700-0430-4cc5-a572-234a2a946748" (UID: "60587700-0430-4cc5-a572-234a2a946748"). InnerVolumeSpecName "kube-api-access-qxkjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.213463 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60587700-0430-4cc5-a572-234a2a946748-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "60587700-0430-4cc5-a572-234a2a946748" (UID: "60587700-0430-4cc5-a572-234a2a946748"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.214783 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60587700-0430-4cc5-a572-234a2a946748-logs" (OuterVolumeSpecName: "logs") pod "60587700-0430-4cc5-a572-234a2a946748" (UID: "60587700-0430-4cc5-a572-234a2a946748"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.216199 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-scripts" (OuterVolumeSpecName: "scripts") pod "60587700-0430-4cc5-a572-234a2a946748" (UID: "60587700-0430-4cc5-a572-234a2a946748"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.224158 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "60587700-0430-4cc5-a572-234a2a946748" (UID: "60587700-0430-4cc5-a572-234a2a946748"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.267141 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60587700-0430-4cc5-a572-234a2a946748" (UID: "60587700-0430-4cc5-a572-234a2a946748"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.309997 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.310027 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.310049 4806 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.310058 4806 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/60587700-0430-4cc5-a572-234a2a946748-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.310068 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60587700-0430-4cc5-a572-234a2a946748-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.310079 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxkjq\" (UniqueName: \"kubernetes.io/projected/60587700-0430-4cc5-a572-234a2a946748-kube-api-access-qxkjq\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.335822 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "60587700-0430-4cc5-a572-234a2a946748" (UID: "60587700-0430-4cc5-a572-234a2a946748"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.362122 4806 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.373402 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-config-data" (OuterVolumeSpecName: "config-data") pod "60587700-0430-4cc5-a572-234a2a946748" (UID: "60587700-0430-4cc5-a572-234a2a946748"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.376074 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.379009 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.411878 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.412138 4806 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.412229 4806 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60587700-0430-4cc5-a572-234a2a946748-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.528632 4806 generic.go:334] "Generic (PLEG): container finished" podID="60587700-0430-4cc5-a572-234a2a946748" containerID="995f3e6bceddd25455c101aa2af9f6e81db185f21e393da766f60b43f9a63f19" exitCode=0 Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.528837 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"60587700-0430-4cc5-a572-234a2a946748","Type":"ContainerDied","Data":"995f3e6bceddd25455c101aa2af9f6e81db185f21e393da766f60b43f9a63f19"} Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.528907 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"60587700-0430-4cc5-a572-234a2a946748","Type":"ContainerDied","Data":"50ea35ebb62b4e8d15ed01641d880bd6f0828594ae397cee91fef8794f4e4425"} Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.528962 4806 scope.go:117] "RemoveContainer" containerID="995f3e6bceddd25455c101aa2af9f6e81db185f21e393da766f60b43f9a63f19" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.529025 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.530201 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.611205 4806 scope.go:117] "RemoveContainer" containerID="0ab275c492bba0348963efaa7a72714cbf7dc07e88ae48f2852a33149e3aa1b2" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.648137 4806 scope.go:117] "RemoveContainer" containerID="995f3e6bceddd25455c101aa2af9f6e81db185f21e393da766f60b43f9a63f19" Dec 04 04:16:10 crc kubenswrapper[4806]: E1204 04:16:10.648668 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"995f3e6bceddd25455c101aa2af9f6e81db185f21e393da766f60b43f9a63f19\": container with ID starting with 995f3e6bceddd25455c101aa2af9f6e81db185f21e393da766f60b43f9a63f19 not found: ID does not exist" containerID="995f3e6bceddd25455c101aa2af9f6e81db185f21e393da766f60b43f9a63f19" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.648730 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"995f3e6bceddd25455c101aa2af9f6e81db185f21e393da766f60b43f9a63f19"} err="failed to get container status \"995f3e6bceddd25455c101aa2af9f6e81db185f21e393da766f60b43f9a63f19\": rpc error: code = NotFound desc = could not find container \"995f3e6bceddd25455c101aa2af9f6e81db185f21e393da766f60b43f9a63f19\": container with ID starting with 995f3e6bceddd25455c101aa2af9f6e81db185f21e393da766f60b43f9a63f19 not found: ID does not exist" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.648777 4806 scope.go:117] "RemoveContainer" containerID="0ab275c492bba0348963efaa7a72714cbf7dc07e88ae48f2852a33149e3aa1b2" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.648898 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 04:16:10 crc kubenswrapper[4806]: E1204 04:16:10.649216 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ab275c492bba0348963efaa7a72714cbf7dc07e88ae48f2852a33149e3aa1b2\": container with ID starting with 0ab275c492bba0348963efaa7a72714cbf7dc07e88ae48f2852a33149e3aa1b2 not found: ID does not exist" containerID="0ab275c492bba0348963efaa7a72714cbf7dc07e88ae48f2852a33149e3aa1b2" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.649240 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ab275c492bba0348963efaa7a72714cbf7dc07e88ae48f2852a33149e3aa1b2"} err="failed to get container status \"0ab275c492bba0348963efaa7a72714cbf7dc07e88ae48f2852a33149e3aa1b2\": rpc error: code = NotFound desc = could not find container \"0ab275c492bba0348963efaa7a72714cbf7dc07e88ae48f2852a33149e3aa1b2\": container with ID starting with 0ab275c492bba0348963efaa7a72714cbf7dc07e88ae48f2852a33149e3aa1b2 not found: ID does not exist" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.669856 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.691392 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.722031 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.730900 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 04:16:10 crc kubenswrapper[4806]: E1204 04:16:10.731521 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06eb71d9-0ccb-4cee-8217-97dae82cbcdf" containerName="glance-httpd" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.731536 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="06eb71d9-0ccb-4cee-8217-97dae82cbcdf" containerName="glance-httpd" Dec 04 04:16:10 crc kubenswrapper[4806]: E1204 04:16:10.731547 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60587700-0430-4cc5-a572-234a2a946748" containerName="glance-httpd" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.731553 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="60587700-0430-4cc5-a572-234a2a946748" containerName="glance-httpd" Dec 04 04:16:10 crc kubenswrapper[4806]: E1204 04:16:10.731569 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60587700-0430-4cc5-a572-234a2a946748" containerName="glance-log" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.731575 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="60587700-0430-4cc5-a572-234a2a946748" containerName="glance-log" Dec 04 04:16:10 crc kubenswrapper[4806]: E1204 04:16:10.731595 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06eb71d9-0ccb-4cee-8217-97dae82cbcdf" containerName="glance-log" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.731601 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="06eb71d9-0ccb-4cee-8217-97dae82cbcdf" containerName="glance-log" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.731784 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="60587700-0430-4cc5-a572-234a2a946748" containerName="glance-log" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.731800 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="06eb71d9-0ccb-4cee-8217-97dae82cbcdf" containerName="glance-log" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.731810 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="06eb71d9-0ccb-4cee-8217-97dae82cbcdf" containerName="glance-httpd" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.731820 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="60587700-0430-4cc5-a572-234a2a946748" containerName="glance-httpd" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.733018 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.738369 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.744126 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-zbhs5" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.744316 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.744426 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.800467 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.845050 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f1138ab-7d74-4dc5-883c-e32dd1f21546-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.845121 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.845150 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f1138ab-7d74-4dc5-883c-e32dd1f21546-logs\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.845193 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f1138ab-7d74-4dc5-883c-e32dd1f21546-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.845232 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f1138ab-7d74-4dc5-883c-e32dd1f21546-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.845289 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f1138ab-7d74-4dc5-883c-e32dd1f21546-scripts\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.845380 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f1138ab-7d74-4dc5-883c-e32dd1f21546-config-data\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.845412 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5bvc\" (UniqueName: \"kubernetes.io/projected/0f1138ab-7d74-4dc5-883c-e32dd1f21546-kube-api-access-c5bvc\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.876377 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.922385 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.931476 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-685c486f6f-6zmmj" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.941000 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.946673 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f1138ab-7d74-4dc5-883c-e32dd1f21546-config-data\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.946724 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5bvc\" (UniqueName: \"kubernetes.io/projected/0f1138ab-7d74-4dc5-883c-e32dd1f21546-kube-api-access-c5bvc\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.946787 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f1138ab-7d74-4dc5-883c-e32dd1f21546-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.946815 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.946852 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f1138ab-7d74-4dc5-883c-e32dd1f21546-logs\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.946879 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f1138ab-7d74-4dc5-883c-e32dd1f21546-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.946918 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f1138ab-7d74-4dc5-883c-e32dd1f21546-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.946977 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f1138ab-7d74-4dc5-883c-e32dd1f21546-scripts\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.947831 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.950396 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f1138ab-7d74-4dc5-883c-e32dd1f21546-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.951363 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f1138ab-7d74-4dc5-883c-e32dd1f21546-logs\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.953050 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.954168 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f1138ab-7d74-4dc5-883c-e32dd1f21546-scripts\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.958231 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.958458 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f1138ab-7d74-4dc5-883c-e32dd1f21546-config-data\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.958478 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.974655 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f1138ab-7d74-4dc5-883c-e32dd1f21546-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:10 crc kubenswrapper[4806]: I1204 04:16:10.978019 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f1138ab-7d74-4dc5-883c-e32dd1f21546-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.094862 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5bvc\" (UniqueName: \"kubernetes.io/projected/0f1138ab-7d74-4dc5-883c-e32dd1f21546-kube-api-access-c5bvc\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.102833 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.103340 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.103377 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.109002 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.109183 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.109504 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlmcf\" (UniqueName: \"kubernetes.io/projected/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-kube-api-access-qlmcf\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.109658 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-logs\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.109790 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.132390 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"0f1138ab-7d74-4dc5-883c-e32dd1f21546\") " pod="openstack/glance-default-external-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.212052 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.213814 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlmcf\" (UniqueName: \"kubernetes.io/projected/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-kube-api-access-qlmcf\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.213871 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-logs\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.213905 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.213981 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.214057 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.214084 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.214146 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.214185 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.214531 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-logs\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.218442 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.218740 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.228437 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.238096 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.250128 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.259022 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlmcf\" (UniqueName: \"kubernetes.io/projected/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-kube-api-access-qlmcf\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.273065 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d515cfb-b26c-430f-9ee9-68b7d1a87fb6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.347002 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6\") " pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.442856 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06eb71d9-0ccb-4cee-8217-97dae82cbcdf" path="/var/lib/kubelet/pods/06eb71d9-0ccb-4cee-8217-97dae82cbcdf/volumes" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.444034 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60587700-0430-4cc5-a572-234a2a946748" path="/var/lib/kubelet/pods/60587700-0430-4cc5-a572-234a2a946748/volumes" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.550097 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"482c5aef-993c-44ea-af1b-60e757b344a1","Type":"ContainerStarted","Data":"642a3088cb61c5436852c2438a8dc0242345c4ba539287cafb7603132511e676"} Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.550487 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="ceilometer-central-agent" containerID="cri-o://773b26bf87027564a689deaa13d0650d55c06518588cdf0ab050f2ab25fe4475" gracePeriod=30 Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.550566 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.550848 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="proxy-httpd" containerID="cri-o://642a3088cb61c5436852c2438a8dc0242345c4ba539287cafb7603132511e676" gracePeriod=30 Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.550894 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="sg-core" containerID="cri-o://2e69d9c7597e4ceef4adc4d6f023fe5bbc10eca0f383bad48be756d435168c3f" gracePeriod=30 Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.550943 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="ceilometer-notification-agent" containerID="cri-o://7ad61da1634012fe21b7a95c9f2df9d3c3c0d4d7ccff7962a606adf6f855e7c3" gracePeriod=30 Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.586114 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.653088319 podStartE2EDuration="7.586097041s" podCreationTimestamp="2025-12-04 04:16:04 +0000 UTC" firstStartedPulling="2025-12-04 04:16:05.747260755 +0000 UTC m=+1280.605773683" lastFinishedPulling="2025-12-04 04:16:10.680269477 +0000 UTC m=+1285.538782405" observedRunningTime="2025-12-04 04:16:11.58449603 +0000 UTC m=+1286.443008958" watchObservedRunningTime="2025-12-04 04:16:11.586097041 +0000 UTC m=+1286.444609969" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.616353 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 04:16:11 crc kubenswrapper[4806]: I1204 04:16:11.958140 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 04:16:11 crc kubenswrapper[4806]: W1204 04:16:11.969589 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f1138ab_7d74_4dc5_883c_e32dd1f21546.slice/crio-eb851b0154daf14653394ce2643fa9eb94e638eea5695aeb82a47701c7318468 WatchSource:0}: Error finding container eb851b0154daf14653394ce2643fa9eb94e638eea5695aeb82a47701c7318468: Status 404 returned error can't find the container with id eb851b0154daf14653394ce2643fa9eb94e638eea5695aeb82a47701c7318468 Dec 04 04:16:12 crc kubenswrapper[4806]: I1204 04:16:12.488035 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 04:16:12 crc kubenswrapper[4806]: W1204 04:16:12.493100 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d515cfb_b26c_430f_9ee9_68b7d1a87fb6.slice/crio-14a8db852b9504dee5a2644e1ac3282e8bb9e023e4f52c0ddb8f3f749bcd9909 WatchSource:0}: Error finding container 14a8db852b9504dee5a2644e1ac3282e8bb9e023e4f52c0ddb8f3f749bcd9909: Status 404 returned error can't find the container with id 14a8db852b9504dee5a2644e1ac3282e8bb9e023e4f52c0ddb8f3f749bcd9909 Dec 04 04:16:12 crc kubenswrapper[4806]: I1204 04:16:12.582339 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f1138ab-7d74-4dc5-883c-e32dd1f21546","Type":"ContainerStarted","Data":"eb851b0154daf14653394ce2643fa9eb94e638eea5695aeb82a47701c7318468"} Dec 04 04:16:12 crc kubenswrapper[4806]: I1204 04:16:12.592416 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6","Type":"ContainerStarted","Data":"14a8db852b9504dee5a2644e1ac3282e8bb9e023e4f52c0ddb8f3f749bcd9909"} Dec 04 04:16:12 crc kubenswrapper[4806]: I1204 04:16:12.596326 4806 generic.go:334] "Generic (PLEG): container finished" podID="482c5aef-993c-44ea-af1b-60e757b344a1" containerID="2e69d9c7597e4ceef4adc4d6f023fe5bbc10eca0f383bad48be756d435168c3f" exitCode=2 Dec 04 04:16:12 crc kubenswrapper[4806]: I1204 04:16:12.596472 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"482c5aef-993c-44ea-af1b-60e757b344a1","Type":"ContainerDied","Data":"2e69d9c7597e4ceef4adc4d6f023fe5bbc10eca0f383bad48be756d435168c3f"} Dec 04 04:16:13 crc kubenswrapper[4806]: I1204 04:16:13.620476 4806 generic.go:334] "Generic (PLEG): container finished" podID="482c5aef-993c-44ea-af1b-60e757b344a1" containerID="7ad61da1634012fe21b7a95c9f2df9d3c3c0d4d7ccff7962a606adf6f855e7c3" exitCode=0 Dec 04 04:16:13 crc kubenswrapper[4806]: I1204 04:16:13.621076 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"482c5aef-993c-44ea-af1b-60e757b344a1","Type":"ContainerDied","Data":"7ad61da1634012fe21b7a95c9f2df9d3c3c0d4d7ccff7962a606adf6f855e7c3"} Dec 04 04:16:13 crc kubenswrapper[4806]: I1204 04:16:13.624388 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f1138ab-7d74-4dc5-883c-e32dd1f21546","Type":"ContainerStarted","Data":"5328095e35cb14669e5f774dceba9dac96594ff2b5546e90871f7f897096c2a9"} Dec 04 04:16:13 crc kubenswrapper[4806]: I1204 04:16:13.630681 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6","Type":"ContainerStarted","Data":"1de852133534bafd9d0d6d77620498f5da3999bbef66ef8a0d31c41a56ac58b0"} Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.393293 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-wqwpc"] Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.394845 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wqwpc" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.460886 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-wqwpc"] Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.511771 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-jt65t"] Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.513083 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jt65t" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.527133 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37d77210-b359-46a2-879c-5422946b6341-operator-scripts\") pod \"nova-api-db-create-wqwpc\" (UID: \"37d77210-b359-46a2-879c-5422946b6341\") " pod="openstack/nova-api-db-create-wqwpc" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.527295 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw9hb\" (UniqueName: \"kubernetes.io/projected/37d77210-b359-46a2-879c-5422946b6341-kube-api-access-bw9hb\") pod \"nova-api-db-create-wqwpc\" (UID: \"37d77210-b359-46a2-879c-5422946b6341\") " pod="openstack/nova-api-db-create-wqwpc" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.545820 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jt65t"] Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.620258 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-61a2-account-create-update-h59lb"] Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.621360 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-61a2-account-create-update-h59lb" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.624528 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.629762 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6d5n\" (UniqueName: \"kubernetes.io/projected/10d9b83f-5515-45d6-86c7-3331dad07c2e-kube-api-access-d6d5n\") pod \"nova-cell0-db-create-jt65t\" (UID: \"10d9b83f-5515-45d6-86c7-3331dad07c2e\") " pod="openstack/nova-cell0-db-create-jt65t" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.629803 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37d77210-b359-46a2-879c-5422946b6341-operator-scripts\") pod \"nova-api-db-create-wqwpc\" (UID: \"37d77210-b359-46a2-879c-5422946b6341\") " pod="openstack/nova-api-db-create-wqwpc" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.629891 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw9hb\" (UniqueName: \"kubernetes.io/projected/37d77210-b359-46a2-879c-5422946b6341-kube-api-access-bw9hb\") pod \"nova-api-db-create-wqwpc\" (UID: \"37d77210-b359-46a2-879c-5422946b6341\") " pod="openstack/nova-api-db-create-wqwpc" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.629940 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10d9b83f-5515-45d6-86c7-3331dad07c2e-operator-scripts\") pod \"nova-cell0-db-create-jt65t\" (UID: \"10d9b83f-5515-45d6-86c7-3331dad07c2e\") " pod="openstack/nova-cell0-db-create-jt65t" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.630635 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37d77210-b359-46a2-879c-5422946b6341-operator-scripts\") pod \"nova-api-db-create-wqwpc\" (UID: \"37d77210-b359-46a2-879c-5422946b6341\") " pod="openstack/nova-api-db-create-wqwpc" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.649872 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5d515cfb-b26c-430f-9ee9-68b7d1a87fb6","Type":"ContainerStarted","Data":"24a27d137190c2ce58370e37aa3767a34a0773377d417316e780c540f4ede939"} Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.653003 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f1138ab-7d74-4dc5-883c-e32dd1f21546","Type":"ContainerStarted","Data":"07a72d02959a596b762ac67bb5b40ef7b74fab95d47ef4ca80cb69de16f80354"} Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.664990 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-61a2-account-create-update-h59lb"] Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.688785 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw9hb\" (UniqueName: \"kubernetes.io/projected/37d77210-b359-46a2-879c-5422946b6341-kube-api-access-bw9hb\") pod \"nova-api-db-create-wqwpc\" (UID: \"37d77210-b359-46a2-879c-5422946b6341\") " pod="openstack/nova-api-db-create-wqwpc" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.731308 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6d5n\" (UniqueName: \"kubernetes.io/projected/10d9b83f-5515-45d6-86c7-3331dad07c2e-kube-api-access-d6d5n\") pod \"nova-cell0-db-create-jt65t\" (UID: \"10d9b83f-5515-45d6-86c7-3331dad07c2e\") " pod="openstack/nova-cell0-db-create-jt65t" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.731367 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86cdc393-070d-423a-869d-bf2c9f5be5f7-operator-scripts\") pod \"nova-api-61a2-account-create-update-h59lb\" (UID: \"86cdc393-070d-423a-869d-bf2c9f5be5f7\") " pod="openstack/nova-api-61a2-account-create-update-h59lb" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.731411 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x5d9\" (UniqueName: \"kubernetes.io/projected/86cdc393-070d-423a-869d-bf2c9f5be5f7-kube-api-access-7x5d9\") pod \"nova-api-61a2-account-create-update-h59lb\" (UID: \"86cdc393-070d-423a-869d-bf2c9f5be5f7\") " pod="openstack/nova-api-61a2-account-create-update-h59lb" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.731466 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10d9b83f-5515-45d6-86c7-3331dad07c2e-operator-scripts\") pod \"nova-cell0-db-create-jt65t\" (UID: \"10d9b83f-5515-45d6-86c7-3331dad07c2e\") " pod="openstack/nova-cell0-db-create-jt65t" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.733354 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10d9b83f-5515-45d6-86c7-3331dad07c2e-operator-scripts\") pod \"nova-cell0-db-create-jt65t\" (UID: \"10d9b83f-5515-45d6-86c7-3331dad07c2e\") " pod="openstack/nova-cell0-db-create-jt65t" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.741867 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.741841672 podStartE2EDuration="4.741841672s" podCreationTimestamp="2025-12-04 04:16:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:16:14.711797964 +0000 UTC m=+1289.570310892" watchObservedRunningTime="2025-12-04 04:16:14.741841672 +0000 UTC m=+1289.600354600" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.761623 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6d5n\" (UniqueName: \"kubernetes.io/projected/10d9b83f-5515-45d6-86c7-3331dad07c2e-kube-api-access-d6d5n\") pod \"nova-cell0-db-create-jt65t\" (UID: \"10d9b83f-5515-45d6-86c7-3331dad07c2e\") " pod="openstack/nova-cell0-db-create-jt65t" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.766114 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wqwpc" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.777008 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-vpg5m"] Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.778279 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vpg5m" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.785177 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-vpg5m"] Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.789281 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.789260219 podStartE2EDuration="4.789260219s" podCreationTimestamp="2025-12-04 04:16:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:16:14.744743463 +0000 UTC m=+1289.603256391" watchObservedRunningTime="2025-12-04 04:16:14.789260219 +0000 UTC m=+1289.647773147" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.830493 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-72b9-account-create-update-pqj5w"] Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.833199 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86cdc393-070d-423a-869d-bf2c9f5be5f7-operator-scripts\") pod \"nova-api-61a2-account-create-update-h59lb\" (UID: \"86cdc393-070d-423a-869d-bf2c9f5be5f7\") " pod="openstack/nova-api-61a2-account-create-update-h59lb" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.833277 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x5d9\" (UniqueName: \"kubernetes.io/projected/86cdc393-070d-423a-869d-bf2c9f5be5f7-kube-api-access-7x5d9\") pod \"nova-api-61a2-account-create-update-h59lb\" (UID: \"86cdc393-070d-423a-869d-bf2c9f5be5f7\") " pod="openstack/nova-api-61a2-account-create-update-h59lb" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.834803 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-72b9-account-create-update-pqj5w" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.835012 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86cdc393-070d-423a-869d-bf2c9f5be5f7-operator-scripts\") pod \"nova-api-61a2-account-create-update-h59lb\" (UID: \"86cdc393-070d-423a-869d-bf2c9f5be5f7\") " pod="openstack/nova-api-61a2-account-create-update-h59lb" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.855692 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jt65t" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.859387 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.892592 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x5d9\" (UniqueName: \"kubernetes.io/projected/86cdc393-070d-423a-869d-bf2c9f5be5f7-kube-api-access-7x5d9\") pod \"nova-api-61a2-account-create-update-h59lb\" (UID: \"86cdc393-070d-423a-869d-bf2c9f5be5f7\") " pod="openstack/nova-api-61a2-account-create-update-h59lb" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.896661 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-72b9-account-create-update-pqj5w"] Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.942185 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-61a2-account-create-update-h59lb" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.953999 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/158f847a-66f8-42c6-b4e0-242dad5a702c-operator-scripts\") pod \"nova-cell1-db-create-vpg5m\" (UID: \"158f847a-66f8-42c6-b4e0-242dad5a702c\") " pod="openstack/nova-cell1-db-create-vpg5m" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.954071 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvkqp\" (UniqueName: \"kubernetes.io/projected/158f847a-66f8-42c6-b4e0-242dad5a702c-kube-api-access-dvkqp\") pod \"nova-cell1-db-create-vpg5m\" (UID: \"158f847a-66f8-42c6-b4e0-242dad5a702c\") " pod="openstack/nova-cell1-db-create-vpg5m" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.954145 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38dd0901-0672-4cac-93d0-881608300b55-operator-scripts\") pod \"nova-cell0-72b9-account-create-update-pqj5w\" (UID: \"38dd0901-0672-4cac-93d0-881608300b55\") " pod="openstack/nova-cell0-72b9-account-create-update-pqj5w" Dec 04 04:16:14 crc kubenswrapper[4806]: I1204 04:16:14.954204 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzzpw\" (UniqueName: \"kubernetes.io/projected/38dd0901-0672-4cac-93d0-881608300b55-kube-api-access-fzzpw\") pod \"nova-cell0-72b9-account-create-update-pqj5w\" (UID: \"38dd0901-0672-4cac-93d0-881608300b55\") " pod="openstack/nova-cell0-72b9-account-create-update-pqj5w" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.013522 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-91c0-account-create-update-2fbzf"] Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.023021 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-91c0-account-create-update-2fbzf" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.036189 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.036599 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-91c0-account-create-update-2fbzf"] Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.063031 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/158f847a-66f8-42c6-b4e0-242dad5a702c-operator-scripts\") pod \"nova-cell1-db-create-vpg5m\" (UID: \"158f847a-66f8-42c6-b4e0-242dad5a702c\") " pod="openstack/nova-cell1-db-create-vpg5m" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.063333 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvkqp\" (UniqueName: \"kubernetes.io/projected/158f847a-66f8-42c6-b4e0-242dad5a702c-kube-api-access-dvkqp\") pod \"nova-cell1-db-create-vpg5m\" (UID: \"158f847a-66f8-42c6-b4e0-242dad5a702c\") " pod="openstack/nova-cell1-db-create-vpg5m" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.063380 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38dd0901-0672-4cac-93d0-881608300b55-operator-scripts\") pod \"nova-cell0-72b9-account-create-update-pqj5w\" (UID: \"38dd0901-0672-4cac-93d0-881608300b55\") " pod="openstack/nova-cell0-72b9-account-create-update-pqj5w" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.063411 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzzpw\" (UniqueName: \"kubernetes.io/projected/38dd0901-0672-4cac-93d0-881608300b55-kube-api-access-fzzpw\") pod \"nova-cell0-72b9-account-create-update-pqj5w\" (UID: \"38dd0901-0672-4cac-93d0-881608300b55\") " pod="openstack/nova-cell0-72b9-account-create-update-pqj5w" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.091021 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/158f847a-66f8-42c6-b4e0-242dad5a702c-operator-scripts\") pod \"nova-cell1-db-create-vpg5m\" (UID: \"158f847a-66f8-42c6-b4e0-242dad5a702c\") " pod="openstack/nova-cell1-db-create-vpg5m" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.091364 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38dd0901-0672-4cac-93d0-881608300b55-operator-scripts\") pod \"nova-cell0-72b9-account-create-update-pqj5w\" (UID: \"38dd0901-0672-4cac-93d0-881608300b55\") " pod="openstack/nova-cell0-72b9-account-create-update-pqj5w" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.122506 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvkqp\" (UniqueName: \"kubernetes.io/projected/158f847a-66f8-42c6-b4e0-242dad5a702c-kube-api-access-dvkqp\") pod \"nova-cell1-db-create-vpg5m\" (UID: \"158f847a-66f8-42c6-b4e0-242dad5a702c\") " pod="openstack/nova-cell1-db-create-vpg5m" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.177327 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzph9\" (UniqueName: \"kubernetes.io/projected/da637fcd-562f-4e71-a068-a4f74b94663f-kube-api-access-hzph9\") pod \"nova-cell1-91c0-account-create-update-2fbzf\" (UID: \"da637fcd-562f-4e71-a068-a4f74b94663f\") " pod="openstack/nova-cell1-91c0-account-create-update-2fbzf" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.177550 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da637fcd-562f-4e71-a068-a4f74b94663f-operator-scripts\") pod \"nova-cell1-91c0-account-create-update-2fbzf\" (UID: \"da637fcd-562f-4e71-a068-a4f74b94663f\") " pod="openstack/nova-cell1-91c0-account-create-update-2fbzf" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.178787 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzzpw\" (UniqueName: \"kubernetes.io/projected/38dd0901-0672-4cac-93d0-881608300b55-kube-api-access-fzzpw\") pod \"nova-cell0-72b9-account-create-update-pqj5w\" (UID: \"38dd0901-0672-4cac-93d0-881608300b55\") " pod="openstack/nova-cell0-72b9-account-create-update-pqj5w" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.254395 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vpg5m" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.274388 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-72b9-account-create-update-pqj5w" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.284039 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da637fcd-562f-4e71-a068-a4f74b94663f-operator-scripts\") pod \"nova-cell1-91c0-account-create-update-2fbzf\" (UID: \"da637fcd-562f-4e71-a068-a4f74b94663f\") " pod="openstack/nova-cell1-91c0-account-create-update-2fbzf" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.284189 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzph9\" (UniqueName: \"kubernetes.io/projected/da637fcd-562f-4e71-a068-a4f74b94663f-kube-api-access-hzph9\") pod \"nova-cell1-91c0-account-create-update-2fbzf\" (UID: \"da637fcd-562f-4e71-a068-a4f74b94663f\") " pod="openstack/nova-cell1-91c0-account-create-update-2fbzf" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.285483 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da637fcd-562f-4e71-a068-a4f74b94663f-operator-scripts\") pod \"nova-cell1-91c0-account-create-update-2fbzf\" (UID: \"da637fcd-562f-4e71-a068-a4f74b94663f\") " pod="openstack/nova-cell1-91c0-account-create-update-2fbzf" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.337441 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzph9\" (UniqueName: \"kubernetes.io/projected/da637fcd-562f-4e71-a068-a4f74b94663f-kube-api-access-hzph9\") pod \"nova-cell1-91c0-account-create-update-2fbzf\" (UID: \"da637fcd-562f-4e71-a068-a4f74b94663f\") " pod="openstack/nova-cell1-91c0-account-create-update-2fbzf" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.395268 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-91c0-account-create-update-2fbzf" Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.809902 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-wqwpc"] Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.823254 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-61a2-account-create-update-h59lb"] Dec 04 04:16:15 crc kubenswrapper[4806]: I1204 04:16:15.863574 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jt65t"] Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.096710 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-72b9-account-create-update-pqj5w"] Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.242889 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-vpg5m"] Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.308643 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-91c0-account-create-update-2fbzf"] Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.707019 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-61a2-account-create-update-h59lb" event={"ID":"86cdc393-070d-423a-869d-bf2c9f5be5f7","Type":"ContainerStarted","Data":"8d6d9ec2c5ccf12239923ee0c0db74aa93fe75dbbaa6b014b81b0a2ab620c83a"} Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.707060 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-61a2-account-create-update-h59lb" event={"ID":"86cdc393-070d-423a-869d-bf2c9f5be5f7","Type":"ContainerStarted","Data":"a589b67b8cdb07dcc19120011a437746021a1fbbba3378537ced3c8e6f363cdd"} Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.719748 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wqwpc" event={"ID":"37d77210-b359-46a2-879c-5422946b6341","Type":"ContainerStarted","Data":"bfadb08507dee6b7a2cb7b6f098d0f50f0840a50e4f3ce650b4728c94302a898"} Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.720109 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wqwpc" event={"ID":"37d77210-b359-46a2-879c-5422946b6341","Type":"ContainerStarted","Data":"33490fd788f90d821bcd4e636b16543f4bfcdd37077cf110c7fd2e2784cd6a9b"} Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.731714 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-72b9-account-create-update-pqj5w" event={"ID":"38dd0901-0672-4cac-93d0-881608300b55","Type":"ContainerStarted","Data":"c69e2e3bc7ed29c334fb3b328c3e2b97610be6079558e730d5a27943dff668e7"} Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.731766 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-72b9-account-create-update-pqj5w" event={"ID":"38dd0901-0672-4cac-93d0-881608300b55","Type":"ContainerStarted","Data":"78fd5a46590222725d9033cfe9a5a975ba057d887733c00ede255a85ccf3cf12"} Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.739870 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jt65t" event={"ID":"10d9b83f-5515-45d6-86c7-3331dad07c2e","Type":"ContainerStarted","Data":"02f52c56412c0e77c78a513b111c8085b365bc749210a2a9e0701d217747edfd"} Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.739938 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jt65t" event={"ID":"10d9b83f-5515-45d6-86c7-3331dad07c2e","Type":"ContainerStarted","Data":"c02d1d0c01f9d7b06ed28f2ef81d634bffdd3c17281d35519659e762c8ba5f35"} Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.755402 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-91c0-account-create-update-2fbzf" event={"ID":"da637fcd-562f-4e71-a068-a4f74b94663f","Type":"ContainerStarted","Data":"73a512b6ad6ee1a5b779dd340db24b0c36c05011f5c66b0b7f1ff54f34dfd47a"} Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.773088 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vpg5m" event={"ID":"158f847a-66f8-42c6-b4e0-242dad5a702c","Type":"ContainerStarted","Data":"91568e3250f24fb58e5225f894168d86d33337304aa52036d0ace330822e95a6"} Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.773138 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vpg5m" event={"ID":"158f847a-66f8-42c6-b4e0-242dad5a702c","Type":"ContainerStarted","Data":"6295d408ee1686a9362da7f52ce09de21515ccf0c8ebfdcde39323d62220f7c3"} Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.783788 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-61a2-account-create-update-h59lb" podStartSLOduration=2.7837644360000002 podStartE2EDuration="2.783764436s" podCreationTimestamp="2025-12-04 04:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:16:16.757474597 +0000 UTC m=+1291.615987525" watchObservedRunningTime="2025-12-04 04:16:16.783764436 +0000 UTC m=+1291.642277364" Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.800236 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-jt65t" podStartSLOduration=2.800216095 podStartE2EDuration="2.800216095s" podCreationTimestamp="2025-12-04 04:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:16:16.786658208 +0000 UTC m=+1291.645171136" watchObservedRunningTime="2025-12-04 04:16:16.800216095 +0000 UTC m=+1291.658729023" Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.812909 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-91c0-account-create-update-2fbzf" podStartSLOduration=2.812886716 podStartE2EDuration="2.812886716s" podCreationTimestamp="2025-12-04 04:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:16:16.812599836 +0000 UTC m=+1291.671112754" watchObservedRunningTime="2025-12-04 04:16:16.812886716 +0000 UTC m=+1291.671399644" Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.850891 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-72b9-account-create-update-pqj5w" podStartSLOduration=2.850867804 podStartE2EDuration="2.850867804s" podCreationTimestamp="2025-12-04 04:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:16:16.835385556 +0000 UTC m=+1291.693898484" watchObservedRunningTime="2025-12-04 04:16:16.850867804 +0000 UTC m=+1291.709380732" Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.874588 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-wqwpc" podStartSLOduration=2.874563552 podStartE2EDuration="2.874563552s" podCreationTimestamp="2025-12-04 04:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:16:16.855331305 +0000 UTC m=+1291.713844233" watchObservedRunningTime="2025-12-04 04:16:16.874563552 +0000 UTC m=+1291.733076480" Dec 04 04:16:16 crc kubenswrapper[4806]: I1204 04:16:16.897400 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-vpg5m" podStartSLOduration=2.897372342 podStartE2EDuration="2.897372342s" podCreationTimestamp="2025-12-04 04:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:16:16.87989678 +0000 UTC m=+1291.738409708" watchObservedRunningTime="2025-12-04 04:16:16.897372342 +0000 UTC m=+1291.755885270" Dec 04 04:16:17 crc kubenswrapper[4806]: I1204 04:16:17.782540 4806 generic.go:334] "Generic (PLEG): container finished" podID="10d9b83f-5515-45d6-86c7-3331dad07c2e" containerID="02f52c56412c0e77c78a513b111c8085b365bc749210a2a9e0701d217747edfd" exitCode=0 Dec 04 04:16:17 crc kubenswrapper[4806]: I1204 04:16:17.782749 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jt65t" event={"ID":"10d9b83f-5515-45d6-86c7-3331dad07c2e","Type":"ContainerDied","Data":"02f52c56412c0e77c78a513b111c8085b365bc749210a2a9e0701d217747edfd"} Dec 04 04:16:17 crc kubenswrapper[4806]: I1204 04:16:17.784712 4806 generic.go:334] "Generic (PLEG): container finished" podID="da637fcd-562f-4e71-a068-a4f74b94663f" containerID="193f0517c4150de6b45ddc9f5267b7220800b401d23ac394dbe24d3b415ab5e5" exitCode=0 Dec 04 04:16:17 crc kubenswrapper[4806]: I1204 04:16:17.784838 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-91c0-account-create-update-2fbzf" event={"ID":"da637fcd-562f-4e71-a068-a4f74b94663f","Type":"ContainerDied","Data":"193f0517c4150de6b45ddc9f5267b7220800b401d23ac394dbe24d3b415ab5e5"} Dec 04 04:16:17 crc kubenswrapper[4806]: I1204 04:16:17.786207 4806 generic.go:334] "Generic (PLEG): container finished" podID="158f847a-66f8-42c6-b4e0-242dad5a702c" containerID="91568e3250f24fb58e5225f894168d86d33337304aa52036d0ace330822e95a6" exitCode=0 Dec 04 04:16:17 crc kubenswrapper[4806]: I1204 04:16:17.786245 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vpg5m" event={"ID":"158f847a-66f8-42c6-b4e0-242dad5a702c","Type":"ContainerDied","Data":"91568e3250f24fb58e5225f894168d86d33337304aa52036d0ace330822e95a6"} Dec 04 04:16:17 crc kubenswrapper[4806]: I1204 04:16:17.787647 4806 generic.go:334] "Generic (PLEG): container finished" podID="86cdc393-070d-423a-869d-bf2c9f5be5f7" containerID="8d6d9ec2c5ccf12239923ee0c0db74aa93fe75dbbaa6b014b81b0a2ab620c83a" exitCode=0 Dec 04 04:16:17 crc kubenswrapper[4806]: I1204 04:16:17.787687 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-61a2-account-create-update-h59lb" event={"ID":"86cdc393-070d-423a-869d-bf2c9f5be5f7","Type":"ContainerDied","Data":"8d6d9ec2c5ccf12239923ee0c0db74aa93fe75dbbaa6b014b81b0a2ab620c83a"} Dec 04 04:16:17 crc kubenswrapper[4806]: I1204 04:16:17.792475 4806 generic.go:334] "Generic (PLEG): container finished" podID="37d77210-b359-46a2-879c-5422946b6341" containerID="bfadb08507dee6b7a2cb7b6f098d0f50f0840a50e4f3ce650b4728c94302a898" exitCode=0 Dec 04 04:16:17 crc kubenswrapper[4806]: I1204 04:16:17.792540 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wqwpc" event={"ID":"37d77210-b359-46a2-879c-5422946b6341","Type":"ContainerDied","Data":"bfadb08507dee6b7a2cb7b6f098d0f50f0840a50e4f3ce650b4728c94302a898"} Dec 04 04:16:17 crc kubenswrapper[4806]: I1204 04:16:17.794721 4806 generic.go:334] "Generic (PLEG): container finished" podID="38dd0901-0672-4cac-93d0-881608300b55" containerID="c69e2e3bc7ed29c334fb3b328c3e2b97610be6079558e730d5a27943dff668e7" exitCode=0 Dec 04 04:16:17 crc kubenswrapper[4806]: I1204 04:16:17.794761 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-72b9-account-create-update-pqj5w" event={"ID":"38dd0901-0672-4cac-93d0-881608300b55","Type":"ContainerDied","Data":"c69e2e3bc7ed29c334fb3b328c3e2b97610be6079558e730d5a27943dff668e7"} Dec 04 04:16:18 crc kubenswrapper[4806]: I1204 04:16:18.807372 4806 generic.go:334] "Generic (PLEG): container finished" podID="482c5aef-993c-44ea-af1b-60e757b344a1" containerID="773b26bf87027564a689deaa13d0650d55c06518588cdf0ab050f2ab25fe4475" exitCode=0 Dec 04 04:16:18 crc kubenswrapper[4806]: I1204 04:16:18.807434 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"482c5aef-993c-44ea-af1b-60e757b344a1","Type":"ContainerDied","Data":"773b26bf87027564a689deaa13d0650d55c06518588cdf0ab050f2ab25fe4475"} Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.372256 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vpg5m" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.444346 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/158f847a-66f8-42c6-b4e0-242dad5a702c-operator-scripts\") pod \"158f847a-66f8-42c6-b4e0-242dad5a702c\" (UID: \"158f847a-66f8-42c6-b4e0-242dad5a702c\") " Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.444651 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvkqp\" (UniqueName: \"kubernetes.io/projected/158f847a-66f8-42c6-b4e0-242dad5a702c-kube-api-access-dvkqp\") pod \"158f847a-66f8-42c6-b4e0-242dad5a702c\" (UID: \"158f847a-66f8-42c6-b4e0-242dad5a702c\") " Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.447368 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/158f847a-66f8-42c6-b4e0-242dad5a702c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "158f847a-66f8-42c6-b4e0-242dad5a702c" (UID: "158f847a-66f8-42c6-b4e0-242dad5a702c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.468366 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/158f847a-66f8-42c6-b4e0-242dad5a702c-kube-api-access-dvkqp" (OuterVolumeSpecName: "kube-api-access-dvkqp") pod "158f847a-66f8-42c6-b4e0-242dad5a702c" (UID: "158f847a-66f8-42c6-b4e0-242dad5a702c"). InnerVolumeSpecName "kube-api-access-dvkqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.556036 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/158f847a-66f8-42c6-b4e0-242dad5a702c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.556081 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvkqp\" (UniqueName: \"kubernetes.io/projected/158f847a-66f8-42c6-b4e0-242dad5a702c-kube-api-access-dvkqp\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.783954 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-61a2-account-create-update-h59lb" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.813917 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-72b9-account-create-update-pqj5w" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.823518 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vpg5m" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.824413 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vpg5m" event={"ID":"158f847a-66f8-42c6-b4e0-242dad5a702c","Type":"ContainerDied","Data":"6295d408ee1686a9362da7f52ce09de21515ccf0c8ebfdcde39323d62220f7c3"} Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.825605 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6295d408ee1686a9362da7f52ce09de21515ccf0c8ebfdcde39323d62220f7c3" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.825750 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wqwpc" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.836303 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-61a2-account-create-update-h59lb" event={"ID":"86cdc393-070d-423a-869d-bf2c9f5be5f7","Type":"ContainerDied","Data":"a589b67b8cdb07dcc19120011a437746021a1fbbba3378537ced3c8e6f363cdd"} Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.836348 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a589b67b8cdb07dcc19120011a437746021a1fbbba3378537ced3c8e6f363cdd" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.836406 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-61a2-account-create-update-h59lb" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.855595 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-72b9-account-create-update-pqj5w" event={"ID":"38dd0901-0672-4cac-93d0-881608300b55","Type":"ContainerDied","Data":"78fd5a46590222725d9033cfe9a5a975ba057d887733c00ede255a85ccf3cf12"} Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.855655 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78fd5a46590222725d9033cfe9a5a975ba057d887733c00ede255a85ccf3cf12" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.855823 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-91c0-account-create-update-2fbzf" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.855920 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-72b9-account-create-update-pqj5w" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.872816 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37d77210-b359-46a2-879c-5422946b6341-operator-scripts\") pod \"37d77210-b359-46a2-879c-5422946b6341\" (UID: \"37d77210-b359-46a2-879c-5422946b6341\") " Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.872904 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86cdc393-070d-423a-869d-bf2c9f5be5f7-operator-scripts\") pod \"86cdc393-070d-423a-869d-bf2c9f5be5f7\" (UID: \"86cdc393-070d-423a-869d-bf2c9f5be5f7\") " Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.872966 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38dd0901-0672-4cac-93d0-881608300b55-operator-scripts\") pod \"38dd0901-0672-4cac-93d0-881608300b55\" (UID: \"38dd0901-0672-4cac-93d0-881608300b55\") " Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.873026 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw9hb\" (UniqueName: \"kubernetes.io/projected/37d77210-b359-46a2-879c-5422946b6341-kube-api-access-bw9hb\") pod \"37d77210-b359-46a2-879c-5422946b6341\" (UID: \"37d77210-b359-46a2-879c-5422946b6341\") " Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.873128 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7x5d9\" (UniqueName: \"kubernetes.io/projected/86cdc393-070d-423a-869d-bf2c9f5be5f7-kube-api-access-7x5d9\") pod \"86cdc393-070d-423a-869d-bf2c9f5be5f7\" (UID: \"86cdc393-070d-423a-869d-bf2c9f5be5f7\") " Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.873223 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzzpw\" (UniqueName: \"kubernetes.io/projected/38dd0901-0672-4cac-93d0-881608300b55-kube-api-access-fzzpw\") pod \"38dd0901-0672-4cac-93d0-881608300b55\" (UID: \"38dd0901-0672-4cac-93d0-881608300b55\") " Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.877092 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38dd0901-0672-4cac-93d0-881608300b55-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "38dd0901-0672-4cac-93d0-881608300b55" (UID: "38dd0901-0672-4cac-93d0-881608300b55"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.877501 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86cdc393-070d-423a-869d-bf2c9f5be5f7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "86cdc393-070d-423a-869d-bf2c9f5be5f7" (UID: "86cdc393-070d-423a-869d-bf2c9f5be5f7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.886187 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wqwpc" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.886731 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wqwpc" event={"ID":"37d77210-b359-46a2-879c-5422946b6341","Type":"ContainerDied","Data":"33490fd788f90d821bcd4e636b16543f4bfcdd37077cf110c7fd2e2784cd6a9b"} Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.886792 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33490fd788f90d821bcd4e636b16543f4bfcdd37077cf110c7fd2e2784cd6a9b" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.892785 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37d77210-b359-46a2-879c-5422946b6341-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "37d77210-b359-46a2-879c-5422946b6341" (UID: "37d77210-b359-46a2-879c-5422946b6341"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.893479 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37d77210-b359-46a2-879c-5422946b6341-kube-api-access-bw9hb" (OuterVolumeSpecName: "kube-api-access-bw9hb") pod "37d77210-b359-46a2-879c-5422946b6341" (UID: "37d77210-b359-46a2-879c-5422946b6341"). InnerVolumeSpecName "kube-api-access-bw9hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.897161 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jt65t" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.902129 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86cdc393-070d-423a-869d-bf2c9f5be5f7-kube-api-access-7x5d9" (OuterVolumeSpecName: "kube-api-access-7x5d9") pod "86cdc393-070d-423a-869d-bf2c9f5be5f7" (UID: "86cdc393-070d-423a-869d-bf2c9f5be5f7"). InnerVolumeSpecName "kube-api-access-7x5d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.902706 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jt65t" event={"ID":"10d9b83f-5515-45d6-86c7-3331dad07c2e","Type":"ContainerDied","Data":"c02d1d0c01f9d7b06ed28f2ef81d634bffdd3c17281d35519659e762c8ba5f35"} Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.904578 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c02d1d0c01f9d7b06ed28f2ef81d634bffdd3c17281d35519659e762c8ba5f35" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.908436 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-91c0-account-create-update-2fbzf" event={"ID":"da637fcd-562f-4e71-a068-a4f74b94663f","Type":"ContainerDied","Data":"73a512b6ad6ee1a5b779dd340db24b0c36c05011f5c66b0b7f1ff54f34dfd47a"} Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.908493 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73a512b6ad6ee1a5b779dd340db24b0c36c05011f5c66b0b7f1ff54f34dfd47a" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.908585 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-91c0-account-create-update-2fbzf" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.914402 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38dd0901-0672-4cac-93d0-881608300b55-kube-api-access-fzzpw" (OuterVolumeSpecName: "kube-api-access-fzzpw") pod "38dd0901-0672-4cac-93d0-881608300b55" (UID: "38dd0901-0672-4cac-93d0-881608300b55"). InnerVolumeSpecName "kube-api-access-fzzpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.957813 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bc4f74d8d-xj2xn" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.979846 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzph9\" (UniqueName: \"kubernetes.io/projected/da637fcd-562f-4e71-a068-a4f74b94663f-kube-api-access-hzph9\") pod \"da637fcd-562f-4e71-a068-a4f74b94663f\" (UID: \"da637fcd-562f-4e71-a068-a4f74b94663f\") " Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.979979 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da637fcd-562f-4e71-a068-a4f74b94663f-operator-scripts\") pod \"da637fcd-562f-4e71-a068-a4f74b94663f\" (UID: \"da637fcd-562f-4e71-a068-a4f74b94663f\") " Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.980030 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10d9b83f-5515-45d6-86c7-3331dad07c2e-operator-scripts\") pod \"10d9b83f-5515-45d6-86c7-3331dad07c2e\" (UID: \"10d9b83f-5515-45d6-86c7-3331dad07c2e\") " Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.980159 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6d5n\" (UniqueName: \"kubernetes.io/projected/10d9b83f-5515-45d6-86c7-3331dad07c2e-kube-api-access-d6d5n\") pod \"10d9b83f-5515-45d6-86c7-3331dad07c2e\" (UID: \"10d9b83f-5515-45d6-86c7-3331dad07c2e\") " Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.981335 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzzpw\" (UniqueName: \"kubernetes.io/projected/38dd0901-0672-4cac-93d0-881608300b55-kube-api-access-fzzpw\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.981376 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37d77210-b359-46a2-879c-5422946b6341-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.981387 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86cdc393-070d-423a-869d-bf2c9f5be5f7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.981396 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38dd0901-0672-4cac-93d0-881608300b55-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.981419 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw9hb\" (UniqueName: \"kubernetes.io/projected/37d77210-b359-46a2-879c-5422946b6341-kube-api-access-bw9hb\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.981429 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7x5d9\" (UniqueName: \"kubernetes.io/projected/86cdc393-070d-423a-869d-bf2c9f5be5f7-kube-api-access-7x5d9\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.983511 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10d9b83f-5515-45d6-86c7-3331dad07c2e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "10d9b83f-5515-45d6-86c7-3331dad07c2e" (UID: "10d9b83f-5515-45d6-86c7-3331dad07c2e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.984348 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da637fcd-562f-4e71-a068-a4f74b94663f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "da637fcd-562f-4e71-a068-a4f74b94663f" (UID: "da637fcd-562f-4e71-a068-a4f74b94663f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.985884 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da637fcd-562f-4e71-a068-a4f74b94663f-kube-api-access-hzph9" (OuterVolumeSpecName: "kube-api-access-hzph9") pod "da637fcd-562f-4e71-a068-a4f74b94663f" (UID: "da637fcd-562f-4e71-a068-a4f74b94663f"). InnerVolumeSpecName "kube-api-access-hzph9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:16:19 crc kubenswrapper[4806]: I1204 04:16:19.990303 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10d9b83f-5515-45d6-86c7-3331dad07c2e-kube-api-access-d6d5n" (OuterVolumeSpecName: "kube-api-access-d6d5n") pod "10d9b83f-5515-45d6-86c7-3331dad07c2e" (UID: "10d9b83f-5515-45d6-86c7-3331dad07c2e"). InnerVolumeSpecName "kube-api-access-d6d5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:16:20 crc kubenswrapper[4806]: I1204 04:16:20.085134 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6d5n\" (UniqueName: \"kubernetes.io/projected/10d9b83f-5515-45d6-86c7-3331dad07c2e-kube-api-access-d6d5n\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:20 crc kubenswrapper[4806]: I1204 04:16:20.085172 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzph9\" (UniqueName: \"kubernetes.io/projected/da637fcd-562f-4e71-a068-a4f74b94663f-kube-api-access-hzph9\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:20 crc kubenswrapper[4806]: I1204 04:16:20.085185 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da637fcd-562f-4e71-a068-a4f74b94663f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:20 crc kubenswrapper[4806]: I1204 04:16:20.085196 4806 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10d9b83f-5515-45d6-86c7-3331dad07c2e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:20 crc kubenswrapper[4806]: I1204 04:16:20.377294 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d8bd6c96d-4sqhf" podUID="d2d53864-c3c8-4e40-8791-0bac49ba561e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 04:16:20 crc kubenswrapper[4806]: I1204 04:16:20.917112 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jt65t" Dec 04 04:16:21 crc kubenswrapper[4806]: I1204 04:16:21.213288 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 04:16:21 crc kubenswrapper[4806]: I1204 04:16:21.215138 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 04:16:21 crc kubenswrapper[4806]: I1204 04:16:21.590956 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 04:16:21 crc kubenswrapper[4806]: I1204 04:16:21.600052 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 04:16:21 crc kubenswrapper[4806]: I1204 04:16:21.617398 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 04:16:21 crc kubenswrapper[4806]: I1204 04:16:21.617447 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 04:16:21 crc kubenswrapper[4806]: I1204 04:16:21.660160 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 04:16:21 crc kubenswrapper[4806]: I1204 04:16:21.675503 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 04:16:21 crc kubenswrapper[4806]: I1204 04:16:21.926849 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 04:16:21 crc kubenswrapper[4806]: I1204 04:16:21.928071 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 04:16:21 crc kubenswrapper[4806]: I1204 04:16:21.928405 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 04:16:21 crc kubenswrapper[4806]: I1204 04:16:21.928438 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 04:16:24 crc kubenswrapper[4806]: I1204 04:16:24.770338 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 04:16:24 crc kubenswrapper[4806]: I1204 04:16:24.770655 4806 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.139801 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.571235 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xcbb8"] Dec 04 04:16:25 crc kubenswrapper[4806]: E1204 04:16:25.571986 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86cdc393-070d-423a-869d-bf2c9f5be5f7" containerName="mariadb-account-create-update" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.572008 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="86cdc393-070d-423a-869d-bf2c9f5be5f7" containerName="mariadb-account-create-update" Dec 04 04:16:25 crc kubenswrapper[4806]: E1204 04:16:25.572039 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10d9b83f-5515-45d6-86c7-3331dad07c2e" containerName="mariadb-database-create" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.572047 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="10d9b83f-5515-45d6-86c7-3331dad07c2e" containerName="mariadb-database-create" Dec 04 04:16:25 crc kubenswrapper[4806]: E1204 04:16:25.572087 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37d77210-b359-46a2-879c-5422946b6341" containerName="mariadb-database-create" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.572096 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="37d77210-b359-46a2-879c-5422946b6341" containerName="mariadb-database-create" Dec 04 04:16:25 crc kubenswrapper[4806]: E1204 04:16:25.572132 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38dd0901-0672-4cac-93d0-881608300b55" containerName="mariadb-account-create-update" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.572141 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="38dd0901-0672-4cac-93d0-881608300b55" containerName="mariadb-account-create-update" Dec 04 04:16:25 crc kubenswrapper[4806]: E1204 04:16:25.572178 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da637fcd-562f-4e71-a068-a4f74b94663f" containerName="mariadb-account-create-update" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.572186 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="da637fcd-562f-4e71-a068-a4f74b94663f" containerName="mariadb-account-create-update" Dec 04 04:16:25 crc kubenswrapper[4806]: E1204 04:16:25.572207 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="158f847a-66f8-42c6-b4e0-242dad5a702c" containerName="mariadb-database-create" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.572216 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="158f847a-66f8-42c6-b4e0-242dad5a702c" containerName="mariadb-database-create" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.572673 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="da637fcd-562f-4e71-a068-a4f74b94663f" containerName="mariadb-account-create-update" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.572705 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="37d77210-b359-46a2-879c-5422946b6341" containerName="mariadb-database-create" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.572718 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="10d9b83f-5515-45d6-86c7-3331dad07c2e" containerName="mariadb-database-create" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.572736 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="158f847a-66f8-42c6-b4e0-242dad5a702c" containerName="mariadb-database-create" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.572766 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="86cdc393-070d-423a-869d-bf2c9f5be5f7" containerName="mariadb-account-create-update" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.572783 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="38dd0901-0672-4cac-93d0-881608300b55" containerName="mariadb-account-create-update" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.583141 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xcbb8" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.589173 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.613814 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fh76g" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.614621 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.638755 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xcbb8"] Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.783371 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xcbb8\" (UID: \"f577bb40-09d5-43ac-ae7b-1e587eed300e\") " pod="openstack/nova-cell0-conductor-db-sync-xcbb8" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.783513 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-scripts\") pod \"nova-cell0-conductor-db-sync-xcbb8\" (UID: \"f577bb40-09d5-43ac-ae7b-1e587eed300e\") " pod="openstack/nova-cell0-conductor-db-sync-xcbb8" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.783556 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-config-data\") pod \"nova-cell0-conductor-db-sync-xcbb8\" (UID: \"f577bb40-09d5-43ac-ae7b-1e587eed300e\") " pod="openstack/nova-cell0-conductor-db-sync-xcbb8" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.783584 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnpgp\" (UniqueName: \"kubernetes.io/projected/f577bb40-09d5-43ac-ae7b-1e587eed300e-kube-api-access-hnpgp\") pod \"nova-cell0-conductor-db-sync-xcbb8\" (UID: \"f577bb40-09d5-43ac-ae7b-1e587eed300e\") " pod="openstack/nova-cell0-conductor-db-sync-xcbb8" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.885267 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-scripts\") pod \"nova-cell0-conductor-db-sync-xcbb8\" (UID: \"f577bb40-09d5-43ac-ae7b-1e587eed300e\") " pod="openstack/nova-cell0-conductor-db-sync-xcbb8" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.885608 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-config-data\") pod \"nova-cell0-conductor-db-sync-xcbb8\" (UID: \"f577bb40-09d5-43ac-ae7b-1e587eed300e\") " pod="openstack/nova-cell0-conductor-db-sync-xcbb8" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.885635 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnpgp\" (UniqueName: \"kubernetes.io/projected/f577bb40-09d5-43ac-ae7b-1e587eed300e-kube-api-access-hnpgp\") pod \"nova-cell0-conductor-db-sync-xcbb8\" (UID: \"f577bb40-09d5-43ac-ae7b-1e587eed300e\") " pod="openstack/nova-cell0-conductor-db-sync-xcbb8" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.885746 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xcbb8\" (UID: \"f577bb40-09d5-43ac-ae7b-1e587eed300e\") " pod="openstack/nova-cell0-conductor-db-sync-xcbb8" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.891509 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-scripts\") pod \"nova-cell0-conductor-db-sync-xcbb8\" (UID: \"f577bb40-09d5-43ac-ae7b-1e587eed300e\") " pod="openstack/nova-cell0-conductor-db-sync-xcbb8" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.891684 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xcbb8\" (UID: \"f577bb40-09d5-43ac-ae7b-1e587eed300e\") " pod="openstack/nova-cell0-conductor-db-sync-xcbb8" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.896794 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-config-data\") pod \"nova-cell0-conductor-db-sync-xcbb8\" (UID: \"f577bb40-09d5-43ac-ae7b-1e587eed300e\") " pod="openstack/nova-cell0-conductor-db-sync-xcbb8" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.914661 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnpgp\" (UniqueName: \"kubernetes.io/projected/f577bb40-09d5-43ac-ae7b-1e587eed300e-kube-api-access-hnpgp\") pod \"nova-cell0-conductor-db-sync-xcbb8\" (UID: \"f577bb40-09d5-43ac-ae7b-1e587eed300e\") " pod="openstack/nova-cell0-conductor-db-sync-xcbb8" Dec 04 04:16:25 crc kubenswrapper[4806]: I1204 04:16:25.933792 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xcbb8" Dec 04 04:16:26 crc kubenswrapper[4806]: I1204 04:16:26.073338 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 04:16:26 crc kubenswrapper[4806]: I1204 04:16:26.073465 4806 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 04:16:26 crc kubenswrapper[4806]: I1204 04:16:26.283858 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 04:16:26 crc kubenswrapper[4806]: I1204 04:16:26.600036 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xcbb8"] Dec 04 04:16:26 crc kubenswrapper[4806]: I1204 04:16:26.629136 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 04:16:26 crc kubenswrapper[4806]: I1204 04:16:26.983348 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xcbb8" event={"ID":"f577bb40-09d5-43ac-ae7b-1e587eed300e","Type":"ContainerStarted","Data":"06599ea39255aca88eceecc66acc9d6ee01f1b6d8e3a901492c17f22e9ad724f"} Dec 04 04:16:27 crc kubenswrapper[4806]: I1204 04:16:27.047611 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:16:27 crc kubenswrapper[4806]: I1204 04:16:27.047678 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:16:27 crc kubenswrapper[4806]: I1204 04:16:27.047732 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 04:16:27 crc kubenswrapper[4806]: I1204 04:16:27.048526 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"95ac262c2400e25c649618e673d25b593b945bc15ad472113e55f250a2d5fdb0"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 04:16:27 crc kubenswrapper[4806]: I1204 04:16:27.048590 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://95ac262c2400e25c649618e673d25b593b945bc15ad472113e55f250a2d5fdb0" gracePeriod=600 Dec 04 04:16:27 crc kubenswrapper[4806]: I1204 04:16:27.999201 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="95ac262c2400e25c649618e673d25b593b945bc15ad472113e55f250a2d5fdb0" exitCode=0 Dec 04 04:16:27 crc kubenswrapper[4806]: I1204 04:16:27.999287 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"95ac262c2400e25c649618e673d25b593b945bc15ad472113e55f250a2d5fdb0"} Dec 04 04:16:27 crc kubenswrapper[4806]: I1204 04:16:27.999514 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc"} Dec 04 04:16:28 crc kubenswrapper[4806]: I1204 04:16:27.999549 4806 scope.go:117] "RemoveContainer" containerID="ff51617027055d5b0c7e244b302f5ed7ba3e3cbe117dc54d0ea82947af1a87f2" Dec 04 04:16:29 crc kubenswrapper[4806]: I1204 04:16:29.944439 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bc4f74d8d-xj2xn" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 04 04:16:30 crc kubenswrapper[4806]: I1204 04:16:30.378076 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d8bd6c96d-4sqhf" podUID="d2d53864-c3c8-4e40-8791-0bac49ba561e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 04:16:35 crc kubenswrapper[4806]: I1204 04:16:35.186947 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 04 04:16:39 crc kubenswrapper[4806]: I1204 04:16:39.944154 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bc4f74d8d-xj2xn" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 04 04:16:39 crc kubenswrapper[4806]: I1204 04:16:39.944657 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:16:39 crc kubenswrapper[4806]: I1204 04:16:39.946024 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"40e0a10f18db6fa40112391f3276df8272d8be2efe99e6d7b0a5ae22be825bb3"} pod="openstack/horizon-7bc4f74d8d-xj2xn" containerMessage="Container horizon failed startup probe, will be restarted" Dec 04 04:16:39 crc kubenswrapper[4806]: I1204 04:16:39.946148 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bc4f74d8d-xj2xn" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" containerID="cri-o://40e0a10f18db6fa40112391f3276df8272d8be2efe99e6d7b0a5ae22be825bb3" gracePeriod=30 Dec 04 04:16:40 crc kubenswrapper[4806]: I1204 04:16:40.377198 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d8bd6c96d-4sqhf" podUID="d2d53864-c3c8-4e40-8791-0bac49ba561e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 04:16:40 crc kubenswrapper[4806]: I1204 04:16:40.377327 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:16:40 crc kubenswrapper[4806]: I1204 04:16:40.378279 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"3d31a9a0fb8d228e6b10c413871aec203543909646c4d72d75fbdde129623505"} pod="openstack/horizon-7d8bd6c96d-4sqhf" containerMessage="Container horizon failed startup probe, will be restarted" Dec 04 04:16:40 crc kubenswrapper[4806]: I1204 04:16:40.378322 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d8bd6c96d-4sqhf" podUID="d2d53864-c3c8-4e40-8791-0bac49ba561e" containerName="horizon" containerID="cri-o://3d31a9a0fb8d228e6b10c413871aec203543909646c4d72d75fbdde129623505" gracePeriod=30 Dec 04 04:16:41 crc kubenswrapper[4806]: I1204 04:16:41.344578 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xcbb8" event={"ID":"f577bb40-09d5-43ac-ae7b-1e587eed300e","Type":"ContainerStarted","Data":"313f402869fe32a6b803febc507ed2f07322acc2418edf9a318d34579cd1eb0c"} Dec 04 04:16:41 crc kubenswrapper[4806]: I1204 04:16:41.374939 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-xcbb8" podStartSLOduration=3.083114714 podStartE2EDuration="16.374906485s" podCreationTimestamp="2025-12-04 04:16:25 +0000 UTC" firstStartedPulling="2025-12-04 04:16:26.628848014 +0000 UTC m=+1301.487360942" lastFinishedPulling="2025-12-04 04:16:39.920639785 +0000 UTC m=+1314.779152713" observedRunningTime="2025-12-04 04:16:41.364984682 +0000 UTC m=+1316.223497620" watchObservedRunningTime="2025-12-04 04:16:41.374906485 +0000 UTC m=+1316.233419423" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.057707 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.139071 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/482c5aef-993c-44ea-af1b-60e757b344a1-log-httpd\") pod \"482c5aef-993c-44ea-af1b-60e757b344a1\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.139165 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-sg-core-conf-yaml\") pod \"482c5aef-993c-44ea-af1b-60e757b344a1\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.139236 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-scripts\") pod \"482c5aef-993c-44ea-af1b-60e757b344a1\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.139292 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-combined-ca-bundle\") pod \"482c5aef-993c-44ea-af1b-60e757b344a1\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.139358 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/482c5aef-993c-44ea-af1b-60e757b344a1-run-httpd\") pod \"482c5aef-993c-44ea-af1b-60e757b344a1\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.139383 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sd2qw\" (UniqueName: \"kubernetes.io/projected/482c5aef-993c-44ea-af1b-60e757b344a1-kube-api-access-sd2qw\") pod \"482c5aef-993c-44ea-af1b-60e757b344a1\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.139410 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-config-data\") pod \"482c5aef-993c-44ea-af1b-60e757b344a1\" (UID: \"482c5aef-993c-44ea-af1b-60e757b344a1\") " Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.148274 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/482c5aef-993c-44ea-af1b-60e757b344a1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "482c5aef-993c-44ea-af1b-60e757b344a1" (UID: "482c5aef-993c-44ea-af1b-60e757b344a1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.154392 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/482c5aef-993c-44ea-af1b-60e757b344a1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "482c5aef-993c-44ea-af1b-60e757b344a1" (UID: "482c5aef-993c-44ea-af1b-60e757b344a1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.154802 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-scripts" (OuterVolumeSpecName: "scripts") pod "482c5aef-993c-44ea-af1b-60e757b344a1" (UID: "482c5aef-993c-44ea-af1b-60e757b344a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.169581 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/482c5aef-993c-44ea-af1b-60e757b344a1-kube-api-access-sd2qw" (OuterVolumeSpecName: "kube-api-access-sd2qw") pod "482c5aef-993c-44ea-af1b-60e757b344a1" (UID: "482c5aef-993c-44ea-af1b-60e757b344a1"). InnerVolumeSpecName "kube-api-access-sd2qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.241123 4806 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/482c5aef-993c-44ea-af1b-60e757b344a1-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.241159 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.241169 4806 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/482c5aef-993c-44ea-af1b-60e757b344a1-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.241178 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sd2qw\" (UniqueName: \"kubernetes.io/projected/482c5aef-993c-44ea-af1b-60e757b344a1-kube-api-access-sd2qw\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.244665 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "482c5aef-993c-44ea-af1b-60e757b344a1" (UID: "482c5aef-993c-44ea-af1b-60e757b344a1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.344067 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "482c5aef-993c-44ea-af1b-60e757b344a1" (UID: "482c5aef-993c-44ea-af1b-60e757b344a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.344082 4806 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.351853 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-config-data" (OuterVolumeSpecName: "config-data") pod "482c5aef-993c-44ea-af1b-60e757b344a1" (UID: "482c5aef-993c-44ea-af1b-60e757b344a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.360646 4806 generic.go:334] "Generic (PLEG): container finished" podID="482c5aef-993c-44ea-af1b-60e757b344a1" containerID="642a3088cb61c5436852c2438a8dc0242345c4ba539287cafb7603132511e676" exitCode=137 Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.360707 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.360722 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"482c5aef-993c-44ea-af1b-60e757b344a1","Type":"ContainerDied","Data":"642a3088cb61c5436852c2438a8dc0242345c4ba539287cafb7603132511e676"} Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.360759 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"482c5aef-993c-44ea-af1b-60e757b344a1","Type":"ContainerDied","Data":"63494d71f85fc8f5e64a02a33ddba2f31064442d8463ae1a86596e0e2c895b7d"} Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.360777 4806 scope.go:117] "RemoveContainer" containerID="642a3088cb61c5436852c2438a8dc0242345c4ba539287cafb7603132511e676" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.393281 4806 scope.go:117] "RemoveContainer" containerID="2e69d9c7597e4ceef4adc4d6f023fe5bbc10eca0f383bad48be756d435168c3f" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.408500 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.438118 4806 scope.go:117] "RemoveContainer" containerID="7ad61da1634012fe21b7a95c9f2df9d3c3c0d4d7ccff7962a606adf6f855e7c3" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.441275 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.445889 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.445916 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/482c5aef-993c-44ea-af1b-60e757b344a1-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.468493 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:16:42 crc kubenswrapper[4806]: E1204 04:16:42.468947 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="proxy-httpd" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.468969 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="proxy-httpd" Dec 04 04:16:42 crc kubenswrapper[4806]: E1204 04:16:42.468990 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="ceilometer-central-agent" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.468999 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="ceilometer-central-agent" Dec 04 04:16:42 crc kubenswrapper[4806]: E1204 04:16:42.469042 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="ceilometer-notification-agent" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.469052 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="ceilometer-notification-agent" Dec 04 04:16:42 crc kubenswrapper[4806]: E1204 04:16:42.469067 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="sg-core" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.469076 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="sg-core" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.469320 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="ceilometer-central-agent" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.469349 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="sg-core" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.469361 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="proxy-httpd" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.469373 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" containerName="ceilometer-notification-agent" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.471307 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.482963 4806 scope.go:117] "RemoveContainer" containerID="773b26bf87027564a689deaa13d0650d55c06518588cdf0ab050f2ab25fe4475" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.483357 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.483573 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.491200 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.525583 4806 scope.go:117] "RemoveContainer" containerID="642a3088cb61c5436852c2438a8dc0242345c4ba539287cafb7603132511e676" Dec 04 04:16:42 crc kubenswrapper[4806]: E1204 04:16:42.526097 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"642a3088cb61c5436852c2438a8dc0242345c4ba539287cafb7603132511e676\": container with ID starting with 642a3088cb61c5436852c2438a8dc0242345c4ba539287cafb7603132511e676 not found: ID does not exist" containerID="642a3088cb61c5436852c2438a8dc0242345c4ba539287cafb7603132511e676" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.526132 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"642a3088cb61c5436852c2438a8dc0242345c4ba539287cafb7603132511e676"} err="failed to get container status \"642a3088cb61c5436852c2438a8dc0242345c4ba539287cafb7603132511e676\": rpc error: code = NotFound desc = could not find container \"642a3088cb61c5436852c2438a8dc0242345c4ba539287cafb7603132511e676\": container with ID starting with 642a3088cb61c5436852c2438a8dc0242345c4ba539287cafb7603132511e676 not found: ID does not exist" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.526153 4806 scope.go:117] "RemoveContainer" containerID="2e69d9c7597e4ceef4adc4d6f023fe5bbc10eca0f383bad48be756d435168c3f" Dec 04 04:16:42 crc kubenswrapper[4806]: E1204 04:16:42.526472 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e69d9c7597e4ceef4adc4d6f023fe5bbc10eca0f383bad48be756d435168c3f\": container with ID starting with 2e69d9c7597e4ceef4adc4d6f023fe5bbc10eca0f383bad48be756d435168c3f not found: ID does not exist" containerID="2e69d9c7597e4ceef4adc4d6f023fe5bbc10eca0f383bad48be756d435168c3f" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.526492 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e69d9c7597e4ceef4adc4d6f023fe5bbc10eca0f383bad48be756d435168c3f"} err="failed to get container status \"2e69d9c7597e4ceef4adc4d6f023fe5bbc10eca0f383bad48be756d435168c3f\": rpc error: code = NotFound desc = could not find container \"2e69d9c7597e4ceef4adc4d6f023fe5bbc10eca0f383bad48be756d435168c3f\": container with ID starting with 2e69d9c7597e4ceef4adc4d6f023fe5bbc10eca0f383bad48be756d435168c3f not found: ID does not exist" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.526507 4806 scope.go:117] "RemoveContainer" containerID="7ad61da1634012fe21b7a95c9f2df9d3c3c0d4d7ccff7962a606adf6f855e7c3" Dec 04 04:16:42 crc kubenswrapper[4806]: E1204 04:16:42.526743 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ad61da1634012fe21b7a95c9f2df9d3c3c0d4d7ccff7962a606adf6f855e7c3\": container with ID starting with 7ad61da1634012fe21b7a95c9f2df9d3c3c0d4d7ccff7962a606adf6f855e7c3 not found: ID does not exist" containerID="7ad61da1634012fe21b7a95c9f2df9d3c3c0d4d7ccff7962a606adf6f855e7c3" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.526765 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ad61da1634012fe21b7a95c9f2df9d3c3c0d4d7ccff7962a606adf6f855e7c3"} err="failed to get container status \"7ad61da1634012fe21b7a95c9f2df9d3c3c0d4d7ccff7962a606adf6f855e7c3\": rpc error: code = NotFound desc = could not find container \"7ad61da1634012fe21b7a95c9f2df9d3c3c0d4d7ccff7962a606adf6f855e7c3\": container with ID starting with 7ad61da1634012fe21b7a95c9f2df9d3c3c0d4d7ccff7962a606adf6f855e7c3 not found: ID does not exist" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.526800 4806 scope.go:117] "RemoveContainer" containerID="773b26bf87027564a689deaa13d0650d55c06518588cdf0ab050f2ab25fe4475" Dec 04 04:16:42 crc kubenswrapper[4806]: E1204 04:16:42.527534 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"773b26bf87027564a689deaa13d0650d55c06518588cdf0ab050f2ab25fe4475\": container with ID starting with 773b26bf87027564a689deaa13d0650d55c06518588cdf0ab050f2ab25fe4475 not found: ID does not exist" containerID="773b26bf87027564a689deaa13d0650d55c06518588cdf0ab050f2ab25fe4475" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.527573 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"773b26bf87027564a689deaa13d0650d55c06518588cdf0ab050f2ab25fe4475"} err="failed to get container status \"773b26bf87027564a689deaa13d0650d55c06518588cdf0ab050f2ab25fe4475\": rpc error: code = NotFound desc = could not find container \"773b26bf87027564a689deaa13d0650d55c06518588cdf0ab050f2ab25fe4475\": container with ID starting with 773b26bf87027564a689deaa13d0650d55c06518588cdf0ab050f2ab25fe4475 not found: ID does not exist" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.547533 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9tqk\" (UniqueName: \"kubernetes.io/projected/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-kube-api-access-s9tqk\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.547655 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-run-httpd\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.547689 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.547798 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.547824 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-config-data\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.547889 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-log-httpd\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.547959 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-scripts\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.649395 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9tqk\" (UniqueName: \"kubernetes.io/projected/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-kube-api-access-s9tqk\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.649481 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-run-httpd\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.649501 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.649565 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.649583 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-config-data\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.649612 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-log-httpd\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.649634 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-scripts\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.657297 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-log-httpd\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.659373 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-run-httpd\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.660383 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.662559 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.667944 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-config-data\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.679226 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-scripts\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.682685 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9tqk\" (UniqueName: \"kubernetes.io/projected/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-kube-api-access-s9tqk\") pod \"ceilometer-0\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " pod="openstack/ceilometer-0" Dec 04 04:16:42 crc kubenswrapper[4806]: I1204 04:16:42.807792 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:16:43 crc kubenswrapper[4806]: I1204 04:16:43.436993 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="482c5aef-993c-44ea-af1b-60e757b344a1" path="/var/lib/kubelet/pods/482c5aef-993c-44ea-af1b-60e757b344a1/volumes" Dec 04 04:16:43 crc kubenswrapper[4806]: W1204 04:16:43.513256 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18170ea9_c07e_4bf4_b6d4_5f72317b3e95.slice/crio-f78e6b536689f2b28dc668a28cf78ec4ac5a5c7054fd36eb792db539b6ed8365 WatchSource:0}: Error finding container f78e6b536689f2b28dc668a28cf78ec4ac5a5c7054fd36eb792db539b6ed8365: Status 404 returned error can't find the container with id f78e6b536689f2b28dc668a28cf78ec4ac5a5c7054fd36eb792db539b6ed8365 Dec 04 04:16:43 crc kubenswrapper[4806]: I1204 04:16:43.515670 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:16:44 crc kubenswrapper[4806]: I1204 04:16:44.407903 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18170ea9-c07e-4bf4-b6d4-5f72317b3e95","Type":"ContainerStarted","Data":"debff5ed08eae6bbd324d0ede5392068b6032a17a9abe84c3d1ddec2d5007110"} Dec 04 04:16:44 crc kubenswrapper[4806]: I1204 04:16:44.408561 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18170ea9-c07e-4bf4-b6d4-5f72317b3e95","Type":"ContainerStarted","Data":"f78e6b536689f2b28dc668a28cf78ec4ac5a5c7054fd36eb792db539b6ed8365"} Dec 04 04:16:45 crc kubenswrapper[4806]: I1204 04:16:45.422133 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18170ea9-c07e-4bf4-b6d4-5f72317b3e95","Type":"ContainerStarted","Data":"06cc56a3b6332c9c218a295b5a11d8b0037b70ee8c4bf7477fe5708cfe28094a"} Dec 04 04:16:46 crc kubenswrapper[4806]: I1204 04:16:46.434880 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18170ea9-c07e-4bf4-b6d4-5f72317b3e95","Type":"ContainerStarted","Data":"e4c78240de1bbe2e3316fc9be8a36d8fb4ca734ae4b31c986e12c0f85331c808"} Dec 04 04:16:47 crc kubenswrapper[4806]: I1204 04:16:47.445760 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18170ea9-c07e-4bf4-b6d4-5f72317b3e95","Type":"ContainerStarted","Data":"a3de0ca96e3c93a15d66d81d98dc2fa8e933d7882004763a2cfcbd34e40d097d"} Dec 04 04:16:47 crc kubenswrapper[4806]: I1204 04:16:47.446143 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 04:16:47 crc kubenswrapper[4806]: I1204 04:16:47.475293 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.032719775 podStartE2EDuration="5.475272526s" podCreationTimestamp="2025-12-04 04:16:42 +0000 UTC" firstStartedPulling="2025-12-04 04:16:43.517578909 +0000 UTC m=+1318.376091837" lastFinishedPulling="2025-12-04 04:16:46.96013164 +0000 UTC m=+1321.818644588" observedRunningTime="2025-12-04 04:16:47.46842537 +0000 UTC m=+1322.326938308" watchObservedRunningTime="2025-12-04 04:16:47.475272526 +0000 UTC m=+1322.333785464" Dec 04 04:16:52 crc kubenswrapper[4806]: I1204 04:16:52.495893 4806 generic.go:334] "Generic (PLEG): container finished" podID="f577bb40-09d5-43ac-ae7b-1e587eed300e" containerID="313f402869fe32a6b803febc507ed2f07322acc2418edf9a318d34579cd1eb0c" exitCode=0 Dec 04 04:16:52 crc kubenswrapper[4806]: I1204 04:16:52.496019 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xcbb8" event={"ID":"f577bb40-09d5-43ac-ae7b-1e587eed300e","Type":"ContainerDied","Data":"313f402869fe32a6b803febc507ed2f07322acc2418edf9a318d34579cd1eb0c"} Dec 04 04:16:53 crc kubenswrapper[4806]: I1204 04:16:53.944303 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xcbb8" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.067752 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnpgp\" (UniqueName: \"kubernetes.io/projected/f577bb40-09d5-43ac-ae7b-1e587eed300e-kube-api-access-hnpgp\") pod \"f577bb40-09d5-43ac-ae7b-1e587eed300e\" (UID: \"f577bb40-09d5-43ac-ae7b-1e587eed300e\") " Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.067818 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-config-data\") pod \"f577bb40-09d5-43ac-ae7b-1e587eed300e\" (UID: \"f577bb40-09d5-43ac-ae7b-1e587eed300e\") " Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.067855 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-combined-ca-bundle\") pod \"f577bb40-09d5-43ac-ae7b-1e587eed300e\" (UID: \"f577bb40-09d5-43ac-ae7b-1e587eed300e\") " Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.067979 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-scripts\") pod \"f577bb40-09d5-43ac-ae7b-1e587eed300e\" (UID: \"f577bb40-09d5-43ac-ae7b-1e587eed300e\") " Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.075371 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-scripts" (OuterVolumeSpecName: "scripts") pod "f577bb40-09d5-43ac-ae7b-1e587eed300e" (UID: "f577bb40-09d5-43ac-ae7b-1e587eed300e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.087353 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f577bb40-09d5-43ac-ae7b-1e587eed300e-kube-api-access-hnpgp" (OuterVolumeSpecName: "kube-api-access-hnpgp") pod "f577bb40-09d5-43ac-ae7b-1e587eed300e" (UID: "f577bb40-09d5-43ac-ae7b-1e587eed300e"). InnerVolumeSpecName "kube-api-access-hnpgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.103014 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-config-data" (OuterVolumeSpecName: "config-data") pod "f577bb40-09d5-43ac-ae7b-1e587eed300e" (UID: "f577bb40-09d5-43ac-ae7b-1e587eed300e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.109312 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f577bb40-09d5-43ac-ae7b-1e587eed300e" (UID: "f577bb40-09d5-43ac-ae7b-1e587eed300e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.170699 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnpgp\" (UniqueName: \"kubernetes.io/projected/f577bb40-09d5-43ac-ae7b-1e587eed300e-kube-api-access-hnpgp\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.171070 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.171084 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.171095 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f577bb40-09d5-43ac-ae7b-1e587eed300e-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.514883 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xcbb8" event={"ID":"f577bb40-09d5-43ac-ae7b-1e587eed300e","Type":"ContainerDied","Data":"06599ea39255aca88eceecc66acc9d6ee01f1b6d8e3a901492c17f22e9ad724f"} Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.514939 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06599ea39255aca88eceecc66acc9d6ee01f1b6d8e3a901492c17f22e9ad724f" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.514993 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xcbb8" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.639047 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 04:16:54 crc kubenswrapper[4806]: E1204 04:16:54.639479 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f577bb40-09d5-43ac-ae7b-1e587eed300e" containerName="nova-cell0-conductor-db-sync" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.639505 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f577bb40-09d5-43ac-ae7b-1e587eed300e" containerName="nova-cell0-conductor-db-sync" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.639733 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f577bb40-09d5-43ac-ae7b-1e587eed300e" containerName="nova-cell0-conductor-db-sync" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.642646 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.649278 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fh76g" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.650005 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.667381 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.680253 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn4s9\" (UniqueName: \"kubernetes.io/projected/713e59da-b517-4099-bf3e-5be89d93e82c-kube-api-access-vn4s9\") pod \"nova-cell0-conductor-0\" (UID: \"713e59da-b517-4099-bf3e-5be89d93e82c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.680307 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/713e59da-b517-4099-bf3e-5be89d93e82c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"713e59da-b517-4099-bf3e-5be89d93e82c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.680339 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/713e59da-b517-4099-bf3e-5be89d93e82c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"713e59da-b517-4099-bf3e-5be89d93e82c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.782280 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn4s9\" (UniqueName: \"kubernetes.io/projected/713e59da-b517-4099-bf3e-5be89d93e82c-kube-api-access-vn4s9\") pod \"nova-cell0-conductor-0\" (UID: \"713e59da-b517-4099-bf3e-5be89d93e82c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.782658 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/713e59da-b517-4099-bf3e-5be89d93e82c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"713e59da-b517-4099-bf3e-5be89d93e82c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.782690 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/713e59da-b517-4099-bf3e-5be89d93e82c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"713e59da-b517-4099-bf3e-5be89d93e82c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.788364 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/713e59da-b517-4099-bf3e-5be89d93e82c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"713e59da-b517-4099-bf3e-5be89d93e82c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.789233 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/713e59da-b517-4099-bf3e-5be89d93e82c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"713e59da-b517-4099-bf3e-5be89d93e82c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.806659 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn4s9\" (UniqueName: \"kubernetes.io/projected/713e59da-b517-4099-bf3e-5be89d93e82c-kube-api-access-vn4s9\") pod \"nova-cell0-conductor-0\" (UID: \"713e59da-b517-4099-bf3e-5be89d93e82c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 04:16:54 crc kubenswrapper[4806]: I1204 04:16:54.963496 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 04:16:55 crc kubenswrapper[4806]: I1204 04:16:55.451820 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 04:16:55 crc kubenswrapper[4806]: I1204 04:16:55.525358 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"713e59da-b517-4099-bf3e-5be89d93e82c","Type":"ContainerStarted","Data":"caf1717afada5a4eca3a3e7b2e1d5ac5f30aca30b5f4fadfecb4eeb4f2f86371"} Dec 04 04:16:56 crc kubenswrapper[4806]: I1204 04:16:56.544118 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"713e59da-b517-4099-bf3e-5be89d93e82c","Type":"ContainerStarted","Data":"dd384c2c253213c8f6504e2c189430a0ad368f3d2a8fbd6341f5299329772672"} Dec 04 04:16:56 crc kubenswrapper[4806]: I1204 04:16:56.544686 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 04 04:17:04 crc kubenswrapper[4806]: I1204 04:17:04.991050 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.006731 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=11.006692031 podStartE2EDuration="11.006692031s" podCreationTimestamp="2025-12-04 04:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:16:56.562974284 +0000 UTC m=+1331.421487212" watchObservedRunningTime="2025-12-04 04:17:05.006692031 +0000 UTC m=+1339.865204959" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.491467 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-zrdpr"] Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.492652 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zrdpr" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.502714 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.504811 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.516363 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zrdpr"] Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.582683 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-scripts\") pod \"nova-cell0-cell-mapping-zrdpr\" (UID: \"79a4caa8-a6ad-47a0-8562-9217d129b010\") " pod="openstack/nova-cell0-cell-mapping-zrdpr" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.582741 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zrdpr\" (UID: \"79a4caa8-a6ad-47a0-8562-9217d129b010\") " pod="openstack/nova-cell0-cell-mapping-zrdpr" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.582841 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj5ds\" (UniqueName: \"kubernetes.io/projected/79a4caa8-a6ad-47a0-8562-9217d129b010-kube-api-access-xj5ds\") pod \"nova-cell0-cell-mapping-zrdpr\" (UID: \"79a4caa8-a6ad-47a0-8562-9217d129b010\") " pod="openstack/nova-cell0-cell-mapping-zrdpr" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.582876 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-config-data\") pod \"nova-cell0-cell-mapping-zrdpr\" (UID: \"79a4caa8-a6ad-47a0-8562-9217d129b010\") " pod="openstack/nova-cell0-cell-mapping-zrdpr" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.640421 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.642421 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.656312 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.661202 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.685800 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj5ds\" (UniqueName: \"kubernetes.io/projected/79a4caa8-a6ad-47a0-8562-9217d129b010-kube-api-access-xj5ds\") pod \"nova-cell0-cell-mapping-zrdpr\" (UID: \"79a4caa8-a6ad-47a0-8562-9217d129b010\") " pod="openstack/nova-cell0-cell-mapping-zrdpr" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.685873 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-config-data\") pod \"nova-cell0-cell-mapping-zrdpr\" (UID: \"79a4caa8-a6ad-47a0-8562-9217d129b010\") " pod="openstack/nova-cell0-cell-mapping-zrdpr" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.686002 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-scripts\") pod \"nova-cell0-cell-mapping-zrdpr\" (UID: \"79a4caa8-a6ad-47a0-8562-9217d129b010\") " pod="openstack/nova-cell0-cell-mapping-zrdpr" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.686030 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f435c60b-81d5-41a0-870a-d2825cc97779-config-data\") pod \"nova-api-0\" (UID: \"f435c60b-81d5-41a0-870a-d2825cc97779\") " pod="openstack/nova-api-0" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.686063 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zrdpr\" (UID: \"79a4caa8-a6ad-47a0-8562-9217d129b010\") " pod="openstack/nova-cell0-cell-mapping-zrdpr" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.686093 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f435c60b-81d5-41a0-870a-d2825cc97779-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f435c60b-81d5-41a0-870a-d2825cc97779\") " pod="openstack/nova-api-0" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.686141 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2mn5\" (UniqueName: \"kubernetes.io/projected/f435c60b-81d5-41a0-870a-d2825cc97779-kube-api-access-p2mn5\") pod \"nova-api-0\" (UID: \"f435c60b-81d5-41a0-870a-d2825cc97779\") " pod="openstack/nova-api-0" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.686167 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f435c60b-81d5-41a0-870a-d2825cc97779-logs\") pod \"nova-api-0\" (UID: \"f435c60b-81d5-41a0-870a-d2825cc97779\") " pod="openstack/nova-api-0" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.696035 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zrdpr\" (UID: \"79a4caa8-a6ad-47a0-8562-9217d129b010\") " pod="openstack/nova-cell0-cell-mapping-zrdpr" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.697021 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-config-data\") pod \"nova-cell0-cell-mapping-zrdpr\" (UID: \"79a4caa8-a6ad-47a0-8562-9217d129b010\") " pod="openstack/nova-cell0-cell-mapping-zrdpr" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.708234 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-scripts\") pod \"nova-cell0-cell-mapping-zrdpr\" (UID: \"79a4caa8-a6ad-47a0-8562-9217d129b010\") " pod="openstack/nova-cell0-cell-mapping-zrdpr" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.720522 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj5ds\" (UniqueName: \"kubernetes.io/projected/79a4caa8-a6ad-47a0-8562-9217d129b010-kube-api-access-xj5ds\") pod \"nova-cell0-cell-mapping-zrdpr\" (UID: \"79a4caa8-a6ad-47a0-8562-9217d129b010\") " pod="openstack/nova-cell0-cell-mapping-zrdpr" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.802535 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f435c60b-81d5-41a0-870a-d2825cc97779-config-data\") pod \"nova-api-0\" (UID: \"f435c60b-81d5-41a0-870a-d2825cc97779\") " pod="openstack/nova-api-0" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.802608 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f435c60b-81d5-41a0-870a-d2825cc97779-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f435c60b-81d5-41a0-870a-d2825cc97779\") " pod="openstack/nova-api-0" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.802651 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2mn5\" (UniqueName: \"kubernetes.io/projected/f435c60b-81d5-41a0-870a-d2825cc97779-kube-api-access-p2mn5\") pod \"nova-api-0\" (UID: \"f435c60b-81d5-41a0-870a-d2825cc97779\") " pod="openstack/nova-api-0" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.802676 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f435c60b-81d5-41a0-870a-d2825cc97779-logs\") pod \"nova-api-0\" (UID: \"f435c60b-81d5-41a0-870a-d2825cc97779\") " pod="openstack/nova-api-0" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.826386 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f435c60b-81d5-41a0-870a-d2825cc97779-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f435c60b-81d5-41a0-870a-d2825cc97779\") " pod="openstack/nova-api-0" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.826599 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zrdpr" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.827404 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f435c60b-81d5-41a0-870a-d2825cc97779-logs\") pod \"nova-api-0\" (UID: \"f435c60b-81d5-41a0-870a-d2825cc97779\") " pod="openstack/nova-api-0" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.836532 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f435c60b-81d5-41a0-870a-d2825cc97779-config-data\") pod \"nova-api-0\" (UID: \"f435c60b-81d5-41a0-870a-d2825cc97779\") " pod="openstack/nova-api-0" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.859799 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.940095 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.948392 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2mn5\" (UniqueName: \"kubernetes.io/projected/f435c60b-81d5-41a0-870a-d2825cc97779-kube-api-access-p2mn5\") pod \"nova-api-0\" (UID: \"f435c60b-81d5-41a0-870a-d2825cc97779\") " pod="openstack/nova-api-0" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.955143 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.977432 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.978874 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.984718 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 04:17:05 crc kubenswrapper[4806]: I1204 04:17:05.986420 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.013790 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.060988 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.140438 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c7c39b0-17d1-4f72-993e-98192c710d83-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6c7c39b0-17d1-4f72-993e-98192c710d83\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.140505 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhbdg\" (UniqueName: \"kubernetes.io/projected/47084642-328d-4596-b54c-07e3ad454765-kube-api-access-rhbdg\") pod \"nova-metadata-0\" (UID: \"47084642-328d-4596-b54c-07e3ad454765\") " pod="openstack/nova-metadata-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.140551 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47084642-328d-4596-b54c-07e3ad454765-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"47084642-328d-4596-b54c-07e3ad454765\") " pod="openstack/nova-metadata-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.140581 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c7c39b0-17d1-4f72-993e-98192c710d83-config-data\") pod \"nova-scheduler-0\" (UID: \"6c7c39b0-17d1-4f72-993e-98192c710d83\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.140607 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47084642-328d-4596-b54c-07e3ad454765-logs\") pod \"nova-metadata-0\" (UID: \"47084642-328d-4596-b54c-07e3ad454765\") " pod="openstack/nova-metadata-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.140630 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47084642-328d-4596-b54c-07e3ad454765-config-data\") pod \"nova-metadata-0\" (UID: \"47084642-328d-4596-b54c-07e3ad454765\") " pod="openstack/nova-metadata-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.140679 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjb7q\" (UniqueName: \"kubernetes.io/projected/6c7c39b0-17d1-4f72-993e-98192c710d83-kube-api-access-jjb7q\") pod \"nova-scheduler-0\" (UID: \"6c7c39b0-17d1-4f72-993e-98192c710d83\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.152966 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-7zc27"] Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.154488 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.201807 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-7zc27"] Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.246416 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c7c39b0-17d1-4f72-993e-98192c710d83-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6c7c39b0-17d1-4f72-993e-98192c710d83\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.246463 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.246495 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-dns-svc\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.246522 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhbdg\" (UniqueName: \"kubernetes.io/projected/47084642-328d-4596-b54c-07e3ad454765-kube-api-access-rhbdg\") pod \"nova-metadata-0\" (UID: \"47084642-328d-4596-b54c-07e3ad454765\") " pod="openstack/nova-metadata-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.246557 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.246575 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-config\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.246589 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47084642-328d-4596-b54c-07e3ad454765-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"47084642-328d-4596-b54c-07e3ad454765\") " pod="openstack/nova-metadata-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.246616 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c7c39b0-17d1-4f72-993e-98192c710d83-config-data\") pod \"nova-scheduler-0\" (UID: \"6c7c39b0-17d1-4f72-993e-98192c710d83\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.246645 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47084642-328d-4596-b54c-07e3ad454765-logs\") pod \"nova-metadata-0\" (UID: \"47084642-328d-4596-b54c-07e3ad454765\") " pod="openstack/nova-metadata-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.246663 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.246685 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47084642-328d-4596-b54c-07e3ad454765-config-data\") pod \"nova-metadata-0\" (UID: \"47084642-328d-4596-b54c-07e3ad454765\") " pod="openstack/nova-metadata-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.246713 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-956ww\" (UniqueName: \"kubernetes.io/projected/733e0561-2895-4aad-a809-40333bf7120d-kube-api-access-956ww\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.246756 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjb7q\" (UniqueName: \"kubernetes.io/projected/6c7c39b0-17d1-4f72-993e-98192c710d83-kube-api-access-jjb7q\") pod \"nova-scheduler-0\" (UID: \"6c7c39b0-17d1-4f72-993e-98192c710d83\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.253342 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47084642-328d-4596-b54c-07e3ad454765-logs\") pod \"nova-metadata-0\" (UID: \"47084642-328d-4596-b54c-07e3ad454765\") " pod="openstack/nova-metadata-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.259582 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c7c39b0-17d1-4f72-993e-98192c710d83-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6c7c39b0-17d1-4f72-993e-98192c710d83\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.261166 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47084642-328d-4596-b54c-07e3ad454765-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"47084642-328d-4596-b54c-07e3ad454765\") " pod="openstack/nova-metadata-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.263591 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47084642-328d-4596-b54c-07e3ad454765-config-data\") pod \"nova-metadata-0\" (UID: \"47084642-328d-4596-b54c-07e3ad454765\") " pod="openstack/nova-metadata-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.272944 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c7c39b0-17d1-4f72-993e-98192c710d83-config-data\") pod \"nova-scheduler-0\" (UID: \"6c7c39b0-17d1-4f72-993e-98192c710d83\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.329741 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjb7q\" (UniqueName: \"kubernetes.io/projected/6c7c39b0-17d1-4f72-993e-98192c710d83-kube-api-access-jjb7q\") pod \"nova-scheduler-0\" (UID: \"6c7c39b0-17d1-4f72-993e-98192c710d83\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.359985 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.361487 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhbdg\" (UniqueName: \"kubernetes.io/projected/47084642-328d-4596-b54c-07e3ad454765-kube-api-access-rhbdg\") pod \"nova-metadata-0\" (UID: \"47084642-328d-4596-b54c-07e3ad454765\") " pod="openstack/nova-metadata-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.361722 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.378288 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.378382 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-dns-svc\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.378493 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.378533 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-config\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.378633 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.378712 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-956ww\" (UniqueName: \"kubernetes.io/projected/733e0561-2895-4aad-a809-40333bf7120d-kube-api-access-956ww\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.379218 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.380045 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.380724 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-dns-svc\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.383136 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.390028 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-config\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.391518 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.393571 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.429287 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-956ww\" (UniqueName: \"kubernetes.io/projected/733e0561-2895-4aad-a809-40333bf7120d-kube-api-access-956ww\") pod \"dnsmasq-dns-757b4f8459-7zc27\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.479958 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.494197 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.494454 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.494588 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nhlt\" (UniqueName: \"kubernetes.io/projected/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-kube-api-access-6nhlt\") pod \"nova-cell1-novncproxy-0\" (UID: \"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.595898 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.596033 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.596075 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nhlt\" (UniqueName: \"kubernetes.io/projected/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-kube-api-access-6nhlt\") pod \"nova-cell1-novncproxy-0\" (UID: \"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.597095 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.603841 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.624578 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.625662 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.627938 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nhlt\" (UniqueName: \"kubernetes.io/projected/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-kube-api-access-6nhlt\") pod \"nova-cell1-novncproxy-0\" (UID: \"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.699513 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:06 crc kubenswrapper[4806]: I1204 04:17:06.718953 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zrdpr"] Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.092861 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:17:07 crc kubenswrapper[4806]: W1204 04:17:07.517461 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47084642_328d_4596_b54c_07e3ad454765.slice/crio-aac4b2b3cb868a9772c25e00915a6d29b89e088706b2c4cfab62ff0711a61e0e WatchSource:0}: Error finding container aac4b2b3cb868a9772c25e00915a6d29b89e088706b2c4cfab62ff0711a61e0e: Status 404 returned error can't find the container with id aac4b2b3cb868a9772c25e00915a6d29b89e088706b2c4cfab62ff0711a61e0e Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.523004 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.563826 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-7zc27"] Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.612539 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7h76s"] Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.615367 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7h76s" Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.618934 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.619132 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.658580 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7h76s"] Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.683262 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.692429 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.752043 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-scripts\") pod \"nova-cell1-conductor-db-sync-7h76s\" (UID: \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\") " pod="openstack/nova-cell1-conductor-db-sync-7h76s" Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.752344 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgwkx\" (UniqueName: \"kubernetes.io/projected/ec8caa98-4401-4f0b-86fa-ff3d04e00544-kube-api-access-zgwkx\") pod \"nova-cell1-conductor-db-sync-7h76s\" (UID: \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\") " pod="openstack/nova-cell1-conductor-db-sync-7h76s" Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.752405 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7h76s\" (UID: \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\") " pod="openstack/nova-cell1-conductor-db-sync-7h76s" Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.752466 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-config-data\") pod \"nova-cell1-conductor-db-sync-7h76s\" (UID: \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\") " pod="openstack/nova-cell1-conductor-db-sync-7h76s" Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.808415 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85","Type":"ContainerStarted","Data":"d2e692527a543c58e8c0faefb4881f1af2ef99cf62c8393270d16f51997df9c6"} Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.820745 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zrdpr" event={"ID":"79a4caa8-a6ad-47a0-8562-9217d129b010","Type":"ContainerStarted","Data":"9981c8f74b5875c56bf20e832a666f1ec7a31accd5a8839c5c0c0b3146dda24c"} Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.820791 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zrdpr" event={"ID":"79a4caa8-a6ad-47a0-8562-9217d129b010","Type":"ContainerStarted","Data":"56b7a88e8d0ce66f43cb1e10d11699542cbc57b72f6637d796797e37d3e56757"} Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.823140 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6c7c39b0-17d1-4f72-993e-98192c710d83","Type":"ContainerStarted","Data":"c4498b068e9da95c8df36bf4775a1a05096d9ace4f0a8c62f50f302602876c1f"} Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.846126 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f435c60b-81d5-41a0-870a-d2825cc97779","Type":"ContainerStarted","Data":"7601b85f1c8dd979d689399aacb8e11ed7320bbe2a7e0244b393aeda9eef73bc"} Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.855129 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgwkx\" (UniqueName: \"kubernetes.io/projected/ec8caa98-4401-4f0b-86fa-ff3d04e00544-kube-api-access-zgwkx\") pod \"nova-cell1-conductor-db-sync-7h76s\" (UID: \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\") " pod="openstack/nova-cell1-conductor-db-sync-7h76s" Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.855217 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7h76s\" (UID: \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\") " pod="openstack/nova-cell1-conductor-db-sync-7h76s" Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.855274 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-config-data\") pod \"nova-cell1-conductor-db-sync-7h76s\" (UID: \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\") " pod="openstack/nova-cell1-conductor-db-sync-7h76s" Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.855433 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-scripts\") pod \"nova-cell1-conductor-db-sync-7h76s\" (UID: \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\") " pod="openstack/nova-cell1-conductor-db-sync-7h76s" Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.860590 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-7zc27" event={"ID":"733e0561-2895-4aad-a809-40333bf7120d","Type":"ContainerStarted","Data":"6801d4478ae4ac246fa054db026e7225daaa33e6c49ee5bf5d2635a24c6da3ab"} Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.869031 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-config-data\") pod \"nova-cell1-conductor-db-sync-7h76s\" (UID: \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\") " pod="openstack/nova-cell1-conductor-db-sync-7h76s" Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.869834 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7h76s\" (UID: \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\") " pod="openstack/nova-cell1-conductor-db-sync-7h76s" Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.870208 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-scripts\") pod \"nova-cell1-conductor-db-sync-7h76s\" (UID: \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\") " pod="openstack/nova-cell1-conductor-db-sync-7h76s" Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.873221 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgwkx\" (UniqueName: \"kubernetes.io/projected/ec8caa98-4401-4f0b-86fa-ff3d04e00544-kube-api-access-zgwkx\") pod \"nova-cell1-conductor-db-sync-7h76s\" (UID: \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\") " pod="openstack/nova-cell1-conductor-db-sync-7h76s" Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.876834 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47084642-328d-4596-b54c-07e3ad454765","Type":"ContainerStarted","Data":"aac4b2b3cb868a9772c25e00915a6d29b89e088706b2c4cfab62ff0711a61e0e"} Dec 04 04:17:07 crc kubenswrapper[4806]: I1204 04:17:07.968593 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7h76s" Dec 04 04:17:08 crc kubenswrapper[4806]: I1204 04:17:08.579544 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-zrdpr" podStartSLOduration=3.579522394 podStartE2EDuration="3.579522394s" podCreationTimestamp="2025-12-04 04:17:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:17:07.855673602 +0000 UTC m=+1342.714186530" watchObservedRunningTime="2025-12-04 04:17:08.579522394 +0000 UTC m=+1343.438035332" Dec 04 04:17:08 crc kubenswrapper[4806]: I1204 04:17:08.588918 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7h76s"] Dec 04 04:17:08 crc kubenswrapper[4806]: I1204 04:17:08.898975 4806 generic.go:334] "Generic (PLEG): container finished" podID="733e0561-2895-4aad-a809-40333bf7120d" containerID="92af52f3015c1fa0dbd735bdf3c206842c8abbeed44a7a6e3efdbded9c3654fe" exitCode=0 Dec 04 04:17:08 crc kubenswrapper[4806]: I1204 04:17:08.899047 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-7zc27" event={"ID":"733e0561-2895-4aad-a809-40333bf7120d","Type":"ContainerDied","Data":"92af52f3015c1fa0dbd735bdf3c206842c8abbeed44a7a6e3efdbded9c3654fe"} Dec 04 04:17:08 crc kubenswrapper[4806]: I1204 04:17:08.905123 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7h76s" event={"ID":"ec8caa98-4401-4f0b-86fa-ff3d04e00544","Type":"ContainerStarted","Data":"b64b7e5c099a47d3dde097dd81dca8b35e560376a0cee7a76e87e2adcf671d41"} Dec 04 04:17:09 crc kubenswrapper[4806]: I1204 04:17:09.927371 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-7zc27" event={"ID":"733e0561-2895-4aad-a809-40333bf7120d","Type":"ContainerStarted","Data":"7fe5631ecec872cd67648b893e68ec39600c044a04eccd576076793dac79abb3"} Dec 04 04:17:09 crc kubenswrapper[4806]: I1204 04:17:09.927748 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:09 crc kubenswrapper[4806]: I1204 04:17:09.931095 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7h76s" event={"ID":"ec8caa98-4401-4f0b-86fa-ff3d04e00544","Type":"ContainerStarted","Data":"060ce8d81fcce975a2212d682de78807111fc2db24634bdc8b95148fa2681ed4"} Dec 04 04:17:09 crc kubenswrapper[4806]: I1204 04:17:09.958668 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-7zc27" podStartSLOduration=3.958654083 podStartE2EDuration="3.958654083s" podCreationTimestamp="2025-12-04 04:17:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:17:09.956705592 +0000 UTC m=+1344.815218520" watchObservedRunningTime="2025-12-04 04:17:09.958654083 +0000 UTC m=+1344.817167011" Dec 04 04:17:09 crc kubenswrapper[4806]: I1204 04:17:09.991275 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-7h76s" podStartSLOduration=2.991256582 podStartE2EDuration="2.991256582s" podCreationTimestamp="2025-12-04 04:17:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:17:09.97945662 +0000 UTC m=+1344.837969548" watchObservedRunningTime="2025-12-04 04:17:09.991256582 +0000 UTC m=+1344.849769510" Dec 04 04:17:10 crc kubenswrapper[4806]: I1204 04:17:10.204989 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:17:10 crc kubenswrapper[4806]: I1204 04:17:10.228304 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 04:17:10 crc kubenswrapper[4806]: I1204 04:17:10.946542 4806 generic.go:334] "Generic (PLEG): container finished" podID="d2d53864-c3c8-4e40-8791-0bac49ba561e" containerID="3d31a9a0fb8d228e6b10c413871aec203543909646c4d72d75fbdde129623505" exitCode=137 Dec 04 04:17:10 crc kubenswrapper[4806]: I1204 04:17:10.946721 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d8bd6c96d-4sqhf" event={"ID":"d2d53864-c3c8-4e40-8791-0bac49ba561e","Type":"ContainerDied","Data":"3d31a9a0fb8d228e6b10c413871aec203543909646c4d72d75fbdde129623505"} Dec 04 04:17:10 crc kubenswrapper[4806]: I1204 04:17:10.947012 4806 scope.go:117] "RemoveContainer" containerID="a440aa262b334c36c507fc4d0fece95b286074ca6be74690b931e0ac6569d085" Dec 04 04:17:10 crc kubenswrapper[4806]: I1204 04:17:10.950241 4806 generic.go:334] "Generic (PLEG): container finished" podID="2265394b-be59-4797-b54d-369e5a5624e0" containerID="40e0a10f18db6fa40112391f3276df8272d8be2efe99e6d7b0a5ae22be825bb3" exitCode=137 Dec 04 04:17:10 crc kubenswrapper[4806]: I1204 04:17:10.950279 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc4f74d8d-xj2xn" event={"ID":"2265394b-be59-4797-b54d-369e5a5624e0","Type":"ContainerDied","Data":"40e0a10f18db6fa40112391f3276df8272d8be2efe99e6d7b0a5ae22be825bb3"} Dec 04 04:17:11 crc kubenswrapper[4806]: I1204 04:17:11.911485 4806 scope.go:117] "RemoveContainer" containerID="33669b61f9a2e337d989305386e72fc8b569e4ce6dbdb13e37c0aacc3aaf1ca1" Dec 04 04:17:12 crc kubenswrapper[4806]: I1204 04:17:12.820668 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 04:17:13 crc kubenswrapper[4806]: I1204 04:17:13.025263 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6c7c39b0-17d1-4f72-993e-98192c710d83","Type":"ContainerStarted","Data":"d08d6e9a7b811235a1a7c85c2244f509e4a26b25d21c0b4b707d4d6e3bb75dbf"} Dec 04 04:17:13 crc kubenswrapper[4806]: I1204 04:17:13.040154 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d8bd6c96d-4sqhf" event={"ID":"d2d53864-c3c8-4e40-8791-0bac49ba561e","Type":"ContainerStarted","Data":"94fb48ffbd3ecbce0978f5312e123b85d6621076e3b49154fec28d0f1c846c75"} Dec 04 04:17:13 crc kubenswrapper[4806]: I1204 04:17:13.066160 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.815551051 podStartE2EDuration="8.066141411s" podCreationTimestamp="2025-12-04 04:17:05 +0000 UTC" firstStartedPulling="2025-12-04 04:17:07.673866335 +0000 UTC m=+1342.532379263" lastFinishedPulling="2025-12-04 04:17:11.924456695 +0000 UTC m=+1346.782969623" observedRunningTime="2025-12-04 04:17:13.055731643 +0000 UTC m=+1347.914244581" watchObservedRunningTime="2025-12-04 04:17:13.066141411 +0000 UTC m=+1347.924654339" Dec 04 04:17:13 crc kubenswrapper[4806]: I1204 04:17:13.080212 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f435c60b-81d5-41a0-870a-d2825cc97779","Type":"ContainerStarted","Data":"3f8c897cb5bfdc9b31b97b8f517383483eb63cf329c035bd24e4c97d97c61359"} Dec 04 04:17:13 crc kubenswrapper[4806]: I1204 04:17:13.080256 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f435c60b-81d5-41a0-870a-d2825cc97779","Type":"ContainerStarted","Data":"e94a9577e68c4802411798687bcaad71b8a6911f920c85a685ad6d906c016e77"} Dec 04 04:17:13 crc kubenswrapper[4806]: I1204 04:17:13.128776 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc4f74d8d-xj2xn" event={"ID":"2265394b-be59-4797-b54d-369e5a5624e0","Type":"ContainerStarted","Data":"93f795e689362ccdc64196f9d3fe1246fe19535562972e9c4bbb63a9dccd7f18"} Dec 04 04:17:13 crc kubenswrapper[4806]: I1204 04:17:13.155753 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47084642-328d-4596-b54c-07e3ad454765","Type":"ContainerStarted","Data":"18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb"} Dec 04 04:17:13 crc kubenswrapper[4806]: I1204 04:17:13.155804 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47084642-328d-4596-b54c-07e3ad454765","Type":"ContainerStarted","Data":"6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0"} Dec 04 04:17:13 crc kubenswrapper[4806]: I1204 04:17:13.155952 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="47084642-328d-4596-b54c-07e3ad454765" containerName="nova-metadata-log" containerID="cri-o://6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0" gracePeriod=30 Dec 04 04:17:13 crc kubenswrapper[4806]: I1204 04:17:13.156364 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="47084642-328d-4596-b54c-07e3ad454765" containerName="nova-metadata-metadata" containerID="cri-o://18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb" gracePeriod=30 Dec 04 04:17:13 crc kubenswrapper[4806]: I1204 04:17:13.159850 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.374547497 podStartE2EDuration="8.159829388s" podCreationTimestamp="2025-12-04 04:17:05 +0000 UTC" firstStartedPulling="2025-12-04 04:17:07.152472093 +0000 UTC m=+1342.010985011" lastFinishedPulling="2025-12-04 04:17:11.937753974 +0000 UTC m=+1346.796266902" observedRunningTime="2025-12-04 04:17:13.109344715 +0000 UTC m=+1347.967857643" watchObservedRunningTime="2025-12-04 04:17:13.159829388 +0000 UTC m=+1348.018342326" Dec 04 04:17:13 crc kubenswrapper[4806]: I1204 04:17:13.162112 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85","Type":"ContainerStarted","Data":"1f3327df49ec06e892dd92691f11321a97476b1e07e1a4b55af59069d2dcf357"} Dec 04 04:17:13 crc kubenswrapper[4806]: I1204 04:17:13.162236 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://1f3327df49ec06e892dd92691f11321a97476b1e07e1a4b55af59069d2dcf357" gracePeriod=30 Dec 04 04:17:13 crc kubenswrapper[4806]: I1204 04:17:13.200492 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.7115946810000002 podStartE2EDuration="8.20047027s" podCreationTimestamp="2025-12-04 04:17:05 +0000 UTC" firstStartedPulling="2025-12-04 04:17:07.51937663 +0000 UTC m=+1342.377889558" lastFinishedPulling="2025-12-04 04:17:12.008252219 +0000 UTC m=+1346.866765147" observedRunningTime="2025-12-04 04:17:13.189454963 +0000 UTC m=+1348.047967901" watchObservedRunningTime="2025-12-04 04:17:13.20047027 +0000 UTC m=+1348.058983198" Dec 04 04:17:13 crc kubenswrapper[4806]: I1204 04:17:13.223781 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.972396642 podStartE2EDuration="7.223760485s" podCreationTimestamp="2025-12-04 04:17:06 +0000 UTC" firstStartedPulling="2025-12-04 04:17:07.669984444 +0000 UTC m=+1342.528497372" lastFinishedPulling="2025-12-04 04:17:11.921348287 +0000 UTC m=+1346.779861215" observedRunningTime="2025-12-04 04:17:13.209824856 +0000 UTC m=+1348.068337784" watchObservedRunningTime="2025-12-04 04:17:13.223760485 +0000 UTC m=+1348.082273413" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.058681 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.139830 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhbdg\" (UniqueName: \"kubernetes.io/projected/47084642-328d-4596-b54c-07e3ad454765-kube-api-access-rhbdg\") pod \"47084642-328d-4596-b54c-07e3ad454765\" (UID: \"47084642-328d-4596-b54c-07e3ad454765\") " Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.140244 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47084642-328d-4596-b54c-07e3ad454765-combined-ca-bundle\") pod \"47084642-328d-4596-b54c-07e3ad454765\" (UID: \"47084642-328d-4596-b54c-07e3ad454765\") " Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.140298 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47084642-328d-4596-b54c-07e3ad454765-config-data\") pod \"47084642-328d-4596-b54c-07e3ad454765\" (UID: \"47084642-328d-4596-b54c-07e3ad454765\") " Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.140330 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47084642-328d-4596-b54c-07e3ad454765-logs\") pod \"47084642-328d-4596-b54c-07e3ad454765\" (UID: \"47084642-328d-4596-b54c-07e3ad454765\") " Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.149088 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47084642-328d-4596-b54c-07e3ad454765-logs" (OuterVolumeSpecName: "logs") pod "47084642-328d-4596-b54c-07e3ad454765" (UID: "47084642-328d-4596-b54c-07e3ad454765"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.195322 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47084642-328d-4596-b54c-07e3ad454765-kube-api-access-rhbdg" (OuterVolumeSpecName: "kube-api-access-rhbdg") pod "47084642-328d-4596-b54c-07e3ad454765" (UID: "47084642-328d-4596-b54c-07e3ad454765"). InnerVolumeSpecName "kube-api-access-rhbdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.223253 4806 generic.go:334] "Generic (PLEG): container finished" podID="47084642-328d-4596-b54c-07e3ad454765" containerID="18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb" exitCode=0 Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.223288 4806 generic.go:334] "Generic (PLEG): container finished" podID="47084642-328d-4596-b54c-07e3ad454765" containerID="6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0" exitCode=143 Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.223319 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.223381 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47084642-328d-4596-b54c-07e3ad454765","Type":"ContainerDied","Data":"18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb"} Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.223413 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47084642-328d-4596-b54c-07e3ad454765","Type":"ContainerDied","Data":"6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0"} Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.223490 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47084642-328d-4596-b54c-07e3ad454765","Type":"ContainerDied","Data":"aac4b2b3cb868a9772c25e00915a6d29b89e088706b2c4cfab62ff0711a61e0e"} Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.223511 4806 scope.go:117] "RemoveContainer" containerID="18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.250579 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhbdg\" (UniqueName: \"kubernetes.io/projected/47084642-328d-4596-b54c-07e3ad454765-kube-api-access-rhbdg\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.250613 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47084642-328d-4596-b54c-07e3ad454765-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.260193 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47084642-328d-4596-b54c-07e3ad454765-config-data" (OuterVolumeSpecName: "config-data") pod "47084642-328d-4596-b54c-07e3ad454765" (UID: "47084642-328d-4596-b54c-07e3ad454765"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.327662 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47084642-328d-4596-b54c-07e3ad454765-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47084642-328d-4596-b54c-07e3ad454765" (UID: "47084642-328d-4596-b54c-07e3ad454765"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.353102 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47084642-328d-4596-b54c-07e3ad454765-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.353152 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47084642-328d-4596-b54c-07e3ad454765-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.376446 4806 scope.go:117] "RemoveContainer" containerID="6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.398614 4806 scope.go:117] "RemoveContainer" containerID="18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb" Dec 04 04:17:14 crc kubenswrapper[4806]: E1204 04:17:14.399091 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb\": container with ID starting with 18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb not found: ID does not exist" containerID="18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.399131 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb"} err="failed to get container status \"18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb\": rpc error: code = NotFound desc = could not find container \"18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb\": container with ID starting with 18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb not found: ID does not exist" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.399157 4806 scope.go:117] "RemoveContainer" containerID="6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0" Dec 04 04:17:14 crc kubenswrapper[4806]: E1204 04:17:14.399416 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0\": container with ID starting with 6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0 not found: ID does not exist" containerID="6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.399442 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0"} err="failed to get container status \"6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0\": rpc error: code = NotFound desc = could not find container \"6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0\": container with ID starting with 6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0 not found: ID does not exist" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.399462 4806 scope.go:117] "RemoveContainer" containerID="18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.399743 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb"} err="failed to get container status \"18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb\": rpc error: code = NotFound desc = could not find container \"18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb\": container with ID starting with 18ac517199297bf5854de96f35fbc3966ffca104375608e3359a9eb48a8606cb not found: ID does not exist" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.399771 4806 scope.go:117] "RemoveContainer" containerID="6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.400033 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0"} err="failed to get container status \"6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0\": rpc error: code = NotFound desc = could not find container \"6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0\": container with ID starting with 6be788be4c44e8f25037da76ee2d4bcae25350561adc9bed3b89d4b05723aef0 not found: ID does not exist" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.566752 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.579691 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.592460 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:17:14 crc kubenswrapper[4806]: E1204 04:17:14.593003 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47084642-328d-4596-b54c-07e3ad454765" containerName="nova-metadata-log" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.593029 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="47084642-328d-4596-b54c-07e3ad454765" containerName="nova-metadata-log" Dec 04 04:17:14 crc kubenswrapper[4806]: E1204 04:17:14.593056 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47084642-328d-4596-b54c-07e3ad454765" containerName="nova-metadata-metadata" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.593065 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="47084642-328d-4596-b54c-07e3ad454765" containerName="nova-metadata-metadata" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.593335 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="47084642-328d-4596-b54c-07e3ad454765" containerName="nova-metadata-metadata" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.593361 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="47084642-328d-4596-b54c-07e3ad454765" containerName="nova-metadata-log" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.594718 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.604453 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.604675 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.623556 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.660050 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.660491 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89d62\" (UniqueName: \"kubernetes.io/projected/ce5df297-871d-4f7b-ac74-616325318f26-kube-api-access-89d62\") pod \"nova-metadata-0\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.660575 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce5df297-871d-4f7b-ac74-616325318f26-logs\") pod \"nova-metadata-0\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.660815 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-config-data\") pod \"nova-metadata-0\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.660846 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.763017 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89d62\" (UniqueName: \"kubernetes.io/projected/ce5df297-871d-4f7b-ac74-616325318f26-kube-api-access-89d62\") pod \"nova-metadata-0\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.763408 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce5df297-871d-4f7b-ac74-616325318f26-logs\") pod \"nova-metadata-0\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.763498 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-config-data\") pod \"nova-metadata-0\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.763526 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.763585 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.764428 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce5df297-871d-4f7b-ac74-616325318f26-logs\") pod \"nova-metadata-0\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.770510 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.770975 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.771101 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-config-data\") pod \"nova-metadata-0\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.788683 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89d62\" (UniqueName: \"kubernetes.io/projected/ce5df297-871d-4f7b-ac74-616325318f26-kube-api-access-89d62\") pod \"nova-metadata-0\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " pod="openstack/nova-metadata-0" Dec 04 04:17:14 crc kubenswrapper[4806]: I1204 04:17:14.920886 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 04:17:15 crc kubenswrapper[4806]: I1204 04:17:15.456344 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47084642-328d-4596-b54c-07e3ad454765" path="/var/lib/kubelet/pods/47084642-328d-4596-b54c-07e3ad454765/volumes" Dec 04 04:17:15 crc kubenswrapper[4806]: I1204 04:17:15.457460 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:17:15 crc kubenswrapper[4806]: I1204 04:17:15.987801 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 04:17:15 crc kubenswrapper[4806]: I1204 04:17:15.988124 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 04:17:16 crc kubenswrapper[4806]: I1204 04:17:16.260218 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ce5df297-871d-4f7b-ac74-616325318f26","Type":"ContainerStarted","Data":"95ed62f0bece6e468c6fbabe37b8c61afc08b48b8c6176bb714a60fd11274810"} Dec 04 04:17:16 crc kubenswrapper[4806]: I1204 04:17:16.260275 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ce5df297-871d-4f7b-ac74-616325318f26","Type":"ContainerStarted","Data":"8c63f8abe1919c41f090e80234924f24282b370031e3e0726293ff3c3921a983"} Dec 04 04:17:16 crc kubenswrapper[4806]: I1204 04:17:16.260292 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ce5df297-871d-4f7b-ac74-616325318f26","Type":"ContainerStarted","Data":"2604b886d8a6b8b11be8477266980fe1aa0a9ea6cff59e8d327aaec4d8afe894"} Dec 04 04:17:16 crc kubenswrapper[4806]: I1204 04:17:16.277447 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.277431866 podStartE2EDuration="2.277431866s" podCreationTimestamp="2025-12-04 04:17:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:17:16.276427854 +0000 UTC m=+1351.134940772" watchObservedRunningTime="2025-12-04 04:17:16.277431866 +0000 UTC m=+1351.135944794" Dec 04 04:17:16 crc kubenswrapper[4806]: I1204 04:17:16.482063 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:16 crc kubenswrapper[4806]: I1204 04:17:16.569541 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5fm5h"] Dec 04 04:17:16 crc kubenswrapper[4806]: I1204 04:17:16.569794 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" podUID="9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb" containerName="dnsmasq-dns" containerID="cri-o://bb221910e129a7ccea0e27f287ad0050a8d2c0dfe3ba999e684e38eadf5d055e" gracePeriod=10 Dec 04 04:17:16 crc kubenswrapper[4806]: I1204 04:17:16.627002 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 04:17:16 crc kubenswrapper[4806]: I1204 04:17:16.627754 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 04:17:16 crc kubenswrapper[4806]: I1204 04:17:16.702021 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:16 crc kubenswrapper[4806]: I1204 04:17:16.710784 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 04:17:17 crc kubenswrapper[4806]: I1204 04:17:17.070092 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f435c60b-81d5-41a0-870a-d2825cc97779" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 04:17:17 crc kubenswrapper[4806]: I1204 04:17:17.070092 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f435c60b-81d5-41a0-870a-d2825cc97779" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 04:17:17 crc kubenswrapper[4806]: I1204 04:17:17.291250 4806 generic.go:334] "Generic (PLEG): container finished" podID="9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb" containerID="bb221910e129a7ccea0e27f287ad0050a8d2c0dfe3ba999e684e38eadf5d055e" exitCode=0 Dec 04 04:17:17 crc kubenswrapper[4806]: I1204 04:17:17.291534 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" event={"ID":"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb","Type":"ContainerDied","Data":"bb221910e129a7ccea0e27f287ad0050a8d2c0dfe3ba999e684e38eadf5d055e"} Dec 04 04:17:17 crc kubenswrapper[4806]: I1204 04:17:17.384359 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 04:17:17 crc kubenswrapper[4806]: I1204 04:17:17.830502 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:17:17 crc kubenswrapper[4806]: I1204 04:17:17.946474 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-dns-svc\") pod \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " Dec 04 04:17:17 crc kubenswrapper[4806]: I1204 04:17:17.946546 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-config\") pod \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " Dec 04 04:17:17 crc kubenswrapper[4806]: I1204 04:17:17.946610 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-dns-swift-storage-0\") pod \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " Dec 04 04:17:17 crc kubenswrapper[4806]: I1204 04:17:17.946634 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-ovsdbserver-sb\") pod \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " Dec 04 04:17:17 crc kubenswrapper[4806]: I1204 04:17:17.946711 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-ovsdbserver-nb\") pod \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " Dec 04 04:17:17 crc kubenswrapper[4806]: I1204 04:17:17.946769 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgjsz\" (UniqueName: \"kubernetes.io/projected/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-kube-api-access-sgjsz\") pod \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\" (UID: \"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb\") " Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.001199 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-kube-api-access-sgjsz" (OuterVolumeSpecName: "kube-api-access-sgjsz") pod "9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb" (UID: "9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb"). InnerVolumeSpecName "kube-api-access-sgjsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.059600 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgjsz\" (UniqueName: \"kubernetes.io/projected/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-kube-api-access-sgjsz\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.150670 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb" (UID: "9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.164593 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.168133 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb" (UID: "9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.181972 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb" (UID: "9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.195728 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-config" (OuterVolumeSpecName: "config") pod "9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb" (UID: "9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.206329 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb" (UID: "9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.266345 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.266392 4806 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.266407 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.266418 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.302814 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.304129 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-5fm5h" event={"ID":"9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb","Type":"ContainerDied","Data":"e600f353e6dc01f6f22a9cdca55a2bff4cf0eb8c33da192ddc6b32c3685fafcb"} Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.304202 4806 scope.go:117] "RemoveContainer" containerID="bb221910e129a7ccea0e27f287ad0050a8d2c0dfe3ba999e684e38eadf5d055e" Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.328371 4806 scope.go:117] "RemoveContainer" containerID="103034db8a4228010d5eba3287cbfba2232e08969f97b8b344d1f9373eec7fa4" Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.359993 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5fm5h"] Dec 04 04:17:18 crc kubenswrapper[4806]: I1204 04:17:18.381287 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-5fm5h"] Dec 04 04:17:19 crc kubenswrapper[4806]: I1204 04:17:19.312579 4806 generic.go:334] "Generic (PLEG): container finished" podID="79a4caa8-a6ad-47a0-8562-9217d129b010" containerID="9981c8f74b5875c56bf20e832a666f1ec7a31accd5a8839c5c0c0b3146dda24c" exitCode=0 Dec 04 04:17:19 crc kubenswrapper[4806]: I1204 04:17:19.312992 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zrdpr" event={"ID":"79a4caa8-a6ad-47a0-8562-9217d129b010","Type":"ContainerDied","Data":"9981c8f74b5875c56bf20e832a666f1ec7a31accd5a8839c5c0c0b3146dda24c"} Dec 04 04:17:19 crc kubenswrapper[4806]: I1204 04:17:19.437473 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb" path="/var/lib/kubelet/pods/9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb/volumes" Dec 04 04:17:19 crc kubenswrapper[4806]: I1204 04:17:19.922181 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 04:17:19 crc kubenswrapper[4806]: I1204 04:17:19.922253 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 04:17:19 crc kubenswrapper[4806]: I1204 04:17:19.943091 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:17:19 crc kubenswrapper[4806]: I1204 04:17:19.943138 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:17:20 crc kubenswrapper[4806]: I1204 04:17:20.071881 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 04:17:20 crc kubenswrapper[4806]: I1204 04:17:20.074195 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="fe13c4d3-85c5-44aa-8424-415f4a775165" containerName="kube-state-metrics" containerID="cri-o://f2ff17d8ba209e64829488fafa2318c3aad6b707b06987d14793cbbc7301c304" gracePeriod=30 Dec 04 04:17:20 crc kubenswrapper[4806]: I1204 04:17:20.328036 4806 generic.go:334] "Generic (PLEG): container finished" podID="fe13c4d3-85c5-44aa-8424-415f4a775165" containerID="f2ff17d8ba209e64829488fafa2318c3aad6b707b06987d14793cbbc7301c304" exitCode=2 Dec 04 04:17:20 crc kubenswrapper[4806]: I1204 04:17:20.328208 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fe13c4d3-85c5-44aa-8424-415f4a775165","Type":"ContainerDied","Data":"f2ff17d8ba209e64829488fafa2318c3aad6b707b06987d14793cbbc7301c304"} Dec 04 04:17:20 crc kubenswrapper[4806]: I1204 04:17:20.376330 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:17:20 crc kubenswrapper[4806]: I1204 04:17:20.376378 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:17:20 crc kubenswrapper[4806]: I1204 04:17:20.631881 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 04:17:20 crc kubenswrapper[4806]: I1204 04:17:20.722995 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlljs\" (UniqueName: \"kubernetes.io/projected/fe13c4d3-85c5-44aa-8424-415f4a775165-kube-api-access-tlljs\") pod \"fe13c4d3-85c5-44aa-8424-415f4a775165\" (UID: \"fe13c4d3-85c5-44aa-8424-415f4a775165\") " Dec 04 04:17:20 crc kubenswrapper[4806]: I1204 04:17:20.764436 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe13c4d3-85c5-44aa-8424-415f4a775165-kube-api-access-tlljs" (OuterVolumeSpecName: "kube-api-access-tlljs") pod "fe13c4d3-85c5-44aa-8424-415f4a775165" (UID: "fe13c4d3-85c5-44aa-8424-415f4a775165"). InnerVolumeSpecName "kube-api-access-tlljs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:17:20 crc kubenswrapper[4806]: I1204 04:17:20.836065 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlljs\" (UniqueName: \"kubernetes.io/projected/fe13c4d3-85c5-44aa-8424-415f4a775165-kube-api-access-tlljs\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:20 crc kubenswrapper[4806]: I1204 04:17:20.873532 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zrdpr" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.039288 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xj5ds\" (UniqueName: \"kubernetes.io/projected/79a4caa8-a6ad-47a0-8562-9217d129b010-kube-api-access-xj5ds\") pod \"79a4caa8-a6ad-47a0-8562-9217d129b010\" (UID: \"79a4caa8-a6ad-47a0-8562-9217d129b010\") " Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.039464 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-config-data\") pod \"79a4caa8-a6ad-47a0-8562-9217d129b010\" (UID: \"79a4caa8-a6ad-47a0-8562-9217d129b010\") " Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.039589 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-scripts\") pod \"79a4caa8-a6ad-47a0-8562-9217d129b010\" (UID: \"79a4caa8-a6ad-47a0-8562-9217d129b010\") " Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.039626 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-combined-ca-bundle\") pod \"79a4caa8-a6ad-47a0-8562-9217d129b010\" (UID: \"79a4caa8-a6ad-47a0-8562-9217d129b010\") " Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.044425 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79a4caa8-a6ad-47a0-8562-9217d129b010-kube-api-access-xj5ds" (OuterVolumeSpecName: "kube-api-access-xj5ds") pod "79a4caa8-a6ad-47a0-8562-9217d129b010" (UID: "79a4caa8-a6ad-47a0-8562-9217d129b010"). InnerVolumeSpecName "kube-api-access-xj5ds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.048076 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-scripts" (OuterVolumeSpecName: "scripts") pod "79a4caa8-a6ad-47a0-8562-9217d129b010" (UID: "79a4caa8-a6ad-47a0-8562-9217d129b010"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.067566 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79a4caa8-a6ad-47a0-8562-9217d129b010" (UID: "79a4caa8-a6ad-47a0-8562-9217d129b010"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.070088 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-config-data" (OuterVolumeSpecName: "config-data") pod "79a4caa8-a6ad-47a0-8562-9217d129b010" (UID: "79a4caa8-a6ad-47a0-8562-9217d129b010"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.141629 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.141674 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.141687 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79a4caa8-a6ad-47a0-8562-9217d129b010-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.141703 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xj5ds\" (UniqueName: \"kubernetes.io/projected/79a4caa8-a6ad-47a0-8562-9217d129b010-kube-api-access-xj5ds\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.338406 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fe13c4d3-85c5-44aa-8424-415f4a775165","Type":"ContainerDied","Data":"619d60dd9550135fd596223448d353f678ea95b7d4dbe4f31b4b7561c63cfd2f"} Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.338458 4806 scope.go:117] "RemoveContainer" containerID="f2ff17d8ba209e64829488fafa2318c3aad6b707b06987d14793cbbc7301c304" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.338571 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.354749 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zrdpr" event={"ID":"79a4caa8-a6ad-47a0-8562-9217d129b010","Type":"ContainerDied","Data":"56b7a88e8d0ce66f43cb1e10d11699542cbc57b72f6637d796797e37d3e56757"} Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.354796 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56b7a88e8d0ce66f43cb1e10d11699542cbc57b72f6637d796797e37d3e56757" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.354868 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zrdpr" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.359741 4806 generic.go:334] "Generic (PLEG): container finished" podID="ec8caa98-4401-4f0b-86fa-ff3d04e00544" containerID="060ce8d81fcce975a2212d682de78807111fc2db24634bdc8b95148fa2681ed4" exitCode=0 Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.359791 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7h76s" event={"ID":"ec8caa98-4401-4f0b-86fa-ff3d04e00544","Type":"ContainerDied","Data":"060ce8d81fcce975a2212d682de78807111fc2db24634bdc8b95148fa2681ed4"} Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.397776 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.417198 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.438287 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe13c4d3-85c5-44aa-8424-415f4a775165" path="/var/lib/kubelet/pods/fe13c4d3-85c5-44aa-8424-415f4a775165/volumes" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.445094 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 04:17:21 crc kubenswrapper[4806]: E1204 04:17:21.445495 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe13c4d3-85c5-44aa-8424-415f4a775165" containerName="kube-state-metrics" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.445513 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe13c4d3-85c5-44aa-8424-415f4a775165" containerName="kube-state-metrics" Dec 04 04:17:21 crc kubenswrapper[4806]: E1204 04:17:21.445539 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79a4caa8-a6ad-47a0-8562-9217d129b010" containerName="nova-manage" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.445547 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="79a4caa8-a6ad-47a0-8562-9217d129b010" containerName="nova-manage" Dec 04 04:17:21 crc kubenswrapper[4806]: E1204 04:17:21.445560 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb" containerName="init" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.445568 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb" containerName="init" Dec 04 04:17:21 crc kubenswrapper[4806]: E1204 04:17:21.445587 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb" containerName="dnsmasq-dns" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.445598 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb" containerName="dnsmasq-dns" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.445858 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="9db2f1c9-7c6e-4d73-957c-a2454bfd2ddb" containerName="dnsmasq-dns" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.445878 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe13c4d3-85c5-44aa-8424-415f4a775165" containerName="kube-state-metrics" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.445898 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="79a4caa8-a6ad-47a0-8562-9217d129b010" containerName="nova-manage" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.446822 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.453087 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.454949 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.457976 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.518614 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.518915 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f435c60b-81d5-41a0-870a-d2825cc97779" containerName="nova-api-log" containerID="cri-o://e94a9577e68c4802411798687bcaad71b8a6911f920c85a685ad6d906c016e77" gracePeriod=30 Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.519059 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f435c60b-81d5-41a0-870a-d2825cc97779" containerName="nova-api-api" containerID="cri-o://3f8c897cb5bfdc9b31b97b8f517383483eb63cf329c035bd24e4c97d97c61359" gracePeriod=30 Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.533990 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.534209 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="6c7c39b0-17d1-4f72-993e-98192c710d83" containerName="nova-scheduler-scheduler" containerID="cri-o://d08d6e9a7b811235a1a7c85c2244f509e4a26b25d21c0b4b707d4d6e3bb75dbf" gracePeriod=30 Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.546807 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.547182 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ce5df297-871d-4f7b-ac74-616325318f26" containerName="nova-metadata-log" containerID="cri-o://8c63f8abe1919c41f090e80234924f24282b370031e3e0726293ff3c3921a983" gracePeriod=30 Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.547330 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ce5df297-871d-4f7b-ac74-616325318f26" containerName="nova-metadata-metadata" containerID="cri-o://95ed62f0bece6e468c6fbabe37b8c61afc08b48b8c6176bb714a60fd11274810" gracePeriod=30 Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.550426 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/00dc1c6d-f6cc-49ac-935f-24007536925c-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"00dc1c6d-f6cc-49ac-935f-24007536925c\") " pod="openstack/kube-state-metrics-0" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.550483 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00dc1c6d-f6cc-49ac-935f-24007536925c-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"00dc1c6d-f6cc-49ac-935f-24007536925c\") " pod="openstack/kube-state-metrics-0" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.550541 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-926qr\" (UniqueName: \"kubernetes.io/projected/00dc1c6d-f6cc-49ac-935f-24007536925c-kube-api-access-926qr\") pod \"kube-state-metrics-0\" (UID: \"00dc1c6d-f6cc-49ac-935f-24007536925c\") " pod="openstack/kube-state-metrics-0" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.550598 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/00dc1c6d-f6cc-49ac-935f-24007536925c-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"00dc1c6d-f6cc-49ac-935f-24007536925c\") " pod="openstack/kube-state-metrics-0" Dec 04 04:17:21 crc kubenswrapper[4806]: E1204 04:17:21.632244 4806 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d08d6e9a7b811235a1a7c85c2244f509e4a26b25d21c0b4b707d4d6e3bb75dbf" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 04:17:21 crc kubenswrapper[4806]: E1204 04:17:21.634138 4806 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d08d6e9a7b811235a1a7c85c2244f509e4a26b25d21c0b4b707d4d6e3bb75dbf" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 04:17:21 crc kubenswrapper[4806]: E1204 04:17:21.637990 4806 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d08d6e9a7b811235a1a7c85c2244f509e4a26b25d21c0b4b707d4d6e3bb75dbf" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 04:17:21 crc kubenswrapper[4806]: E1204 04:17:21.638178 4806 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="6c7c39b0-17d1-4f72-993e-98192c710d83" containerName="nova-scheduler-scheduler" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.652909 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/00dc1c6d-f6cc-49ac-935f-24007536925c-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"00dc1c6d-f6cc-49ac-935f-24007536925c\") " pod="openstack/kube-state-metrics-0" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.653184 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00dc1c6d-f6cc-49ac-935f-24007536925c-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"00dc1c6d-f6cc-49ac-935f-24007536925c\") " pod="openstack/kube-state-metrics-0" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.653334 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-926qr\" (UniqueName: \"kubernetes.io/projected/00dc1c6d-f6cc-49ac-935f-24007536925c-kube-api-access-926qr\") pod \"kube-state-metrics-0\" (UID: \"00dc1c6d-f6cc-49ac-935f-24007536925c\") " pod="openstack/kube-state-metrics-0" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.654306 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/00dc1c6d-f6cc-49ac-935f-24007536925c-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"00dc1c6d-f6cc-49ac-935f-24007536925c\") " pod="openstack/kube-state-metrics-0" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.659514 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00dc1c6d-f6cc-49ac-935f-24007536925c-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"00dc1c6d-f6cc-49ac-935f-24007536925c\") " pod="openstack/kube-state-metrics-0" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.660132 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/00dc1c6d-f6cc-49ac-935f-24007536925c-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"00dc1c6d-f6cc-49ac-935f-24007536925c\") " pod="openstack/kube-state-metrics-0" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.661127 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/00dc1c6d-f6cc-49ac-935f-24007536925c-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"00dc1c6d-f6cc-49ac-935f-24007536925c\") " pod="openstack/kube-state-metrics-0" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.686914 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-926qr\" (UniqueName: \"kubernetes.io/projected/00dc1c6d-f6cc-49ac-935f-24007536925c-kube-api-access-926qr\") pod \"kube-state-metrics-0\" (UID: \"00dc1c6d-f6cc-49ac-935f-24007536925c\") " pod="openstack/kube-state-metrics-0" Dec 04 04:17:21 crc kubenswrapper[4806]: I1204 04:17:21.765663 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.377495 4806 generic.go:334] "Generic (PLEG): container finished" podID="ce5df297-871d-4f7b-ac74-616325318f26" containerID="95ed62f0bece6e468c6fbabe37b8c61afc08b48b8c6176bb714a60fd11274810" exitCode=0 Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.377777 4806 generic.go:334] "Generic (PLEG): container finished" podID="ce5df297-871d-4f7b-ac74-616325318f26" containerID="8c63f8abe1919c41f090e80234924f24282b370031e3e0726293ff3c3921a983" exitCode=143 Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.377819 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ce5df297-871d-4f7b-ac74-616325318f26","Type":"ContainerDied","Data":"95ed62f0bece6e468c6fbabe37b8c61afc08b48b8c6176bb714a60fd11274810"} Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.377844 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ce5df297-871d-4f7b-ac74-616325318f26","Type":"ContainerDied","Data":"8c63f8abe1919c41f090e80234924f24282b370031e3e0726293ff3c3921a983"} Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.377853 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ce5df297-871d-4f7b-ac74-616325318f26","Type":"ContainerDied","Data":"2604b886d8a6b8b11be8477266980fe1aa0a9ea6cff59e8d327aaec4d8afe894"} Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.377862 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2604b886d8a6b8b11be8477266980fe1aa0a9ea6cff59e8d327aaec4d8afe894" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.379983 4806 generic.go:334] "Generic (PLEG): container finished" podID="f435c60b-81d5-41a0-870a-d2825cc97779" containerID="e94a9577e68c4802411798687bcaad71b8a6911f920c85a685ad6d906c016e77" exitCode=143 Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.380029 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f435c60b-81d5-41a0-870a-d2825cc97779","Type":"ContainerDied","Data":"e94a9577e68c4802411798687bcaad71b8a6911f920c85a685ad6d906c016e77"} Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.462770 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.505028 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.705813 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-config-data\") pod \"ce5df297-871d-4f7b-ac74-616325318f26\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.705908 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-combined-ca-bundle\") pod \"ce5df297-871d-4f7b-ac74-616325318f26\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.706025 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-nova-metadata-tls-certs\") pod \"ce5df297-871d-4f7b-ac74-616325318f26\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.706094 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce5df297-871d-4f7b-ac74-616325318f26-logs\") pod \"ce5df297-871d-4f7b-ac74-616325318f26\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.706125 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89d62\" (UniqueName: \"kubernetes.io/projected/ce5df297-871d-4f7b-ac74-616325318f26-kube-api-access-89d62\") pod \"ce5df297-871d-4f7b-ac74-616325318f26\" (UID: \"ce5df297-871d-4f7b-ac74-616325318f26\") " Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.707363 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce5df297-871d-4f7b-ac74-616325318f26-logs" (OuterVolumeSpecName: "logs") pod "ce5df297-871d-4f7b-ac74-616325318f26" (UID: "ce5df297-871d-4f7b-ac74-616325318f26"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.747532 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce5df297-871d-4f7b-ac74-616325318f26-kube-api-access-89d62" (OuterVolumeSpecName: "kube-api-access-89d62") pod "ce5df297-871d-4f7b-ac74-616325318f26" (UID: "ce5df297-871d-4f7b-ac74-616325318f26"). InnerVolumeSpecName "kube-api-access-89d62". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.762637 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-config-data" (OuterVolumeSpecName: "config-data") pod "ce5df297-871d-4f7b-ac74-616325318f26" (UID: "ce5df297-871d-4f7b-ac74-616325318f26"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.794061 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce5df297-871d-4f7b-ac74-616325318f26" (UID: "ce5df297-871d-4f7b-ac74-616325318f26"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.811560 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce5df297-871d-4f7b-ac74-616325318f26-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.811590 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89d62\" (UniqueName: \"kubernetes.io/projected/ce5df297-871d-4f7b-ac74-616325318f26-kube-api-access-89d62\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.811599 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.811610 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.856077 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "ce5df297-871d-4f7b-ac74-616325318f26" (UID: "ce5df297-871d-4f7b-ac74-616325318f26"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.859457 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7h76s" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.913907 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-scripts\") pod \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\" (UID: \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\") " Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.914042 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgwkx\" (UniqueName: \"kubernetes.io/projected/ec8caa98-4401-4f0b-86fa-ff3d04e00544-kube-api-access-zgwkx\") pod \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\" (UID: \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\") " Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.914077 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-combined-ca-bundle\") pod \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\" (UID: \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\") " Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.914155 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-config-data\") pod \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\" (UID: \"ec8caa98-4401-4f0b-86fa-ff3d04e00544\") " Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.914436 4806 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce5df297-871d-4f7b-ac74-616325318f26-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.918912 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-scripts" (OuterVolumeSpecName: "scripts") pod "ec8caa98-4401-4f0b-86fa-ff3d04e00544" (UID: "ec8caa98-4401-4f0b-86fa-ff3d04e00544"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.922705 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec8caa98-4401-4f0b-86fa-ff3d04e00544-kube-api-access-zgwkx" (OuterVolumeSpecName: "kube-api-access-zgwkx") pod "ec8caa98-4401-4f0b-86fa-ff3d04e00544" (UID: "ec8caa98-4401-4f0b-86fa-ff3d04e00544"). InnerVolumeSpecName "kube-api-access-zgwkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.948159 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-config-data" (OuterVolumeSpecName: "config-data") pod "ec8caa98-4401-4f0b-86fa-ff3d04e00544" (UID: "ec8caa98-4401-4f0b-86fa-ff3d04e00544"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:22 crc kubenswrapper[4806]: I1204 04:17:22.956084 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec8caa98-4401-4f0b-86fa-ff3d04e00544" (UID: "ec8caa98-4401-4f0b-86fa-ff3d04e00544"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.015436 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.015481 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgwkx\" (UniqueName: \"kubernetes.io/projected/ec8caa98-4401-4f0b-86fa-ff3d04e00544-kube-api-access-zgwkx\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.015496 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.015509 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8caa98-4401-4f0b-86fa-ff3d04e00544-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.429378 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"00dc1c6d-f6cc-49ac-935f-24007536925c","Type":"ContainerStarted","Data":"2ba6c4245cc5f7ade6bd4a467533e29ceae3faf465f99c7fb1b461c802f1744e"} Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.430076 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"00dc1c6d-f6cc-49ac-935f-24007536925c","Type":"ContainerStarted","Data":"a6cb040f70b374b44dbbe5c106fba56aa79f2791ffe7100081e8a2d5fc38b04d"} Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.430127 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.438012 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.438380 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7h76s" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.476824 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.020326503 podStartE2EDuration="2.476802517s" podCreationTimestamp="2025-12-04 04:17:21 +0000 UTC" firstStartedPulling="2025-12-04 04:17:22.478723152 +0000 UTC m=+1357.337236080" lastFinishedPulling="2025-12-04 04:17:22.935199166 +0000 UTC m=+1357.793712094" observedRunningTime="2025-12-04 04:17:23.465610794 +0000 UTC m=+1358.324123722" watchObservedRunningTime="2025-12-04 04:17:23.476802517 +0000 UTC m=+1358.335315445" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.546031 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7h76s" event={"ID":"ec8caa98-4401-4f0b-86fa-ff3d04e00544","Type":"ContainerDied","Data":"b64b7e5c099a47d3dde097dd81dca8b35e560376a0cee7a76e87e2adcf671d41"} Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.546083 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b64b7e5c099a47d3dde097dd81dca8b35e560376a0cee7a76e87e2adcf671d41" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.564361 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 04:17:23 crc kubenswrapper[4806]: E1204 04:17:23.564857 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec8caa98-4401-4f0b-86fa-ff3d04e00544" containerName="nova-cell1-conductor-db-sync" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.564872 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec8caa98-4401-4f0b-86fa-ff3d04e00544" containerName="nova-cell1-conductor-db-sync" Dec 04 04:17:23 crc kubenswrapper[4806]: E1204 04:17:23.565040 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5df297-871d-4f7b-ac74-616325318f26" containerName="nova-metadata-log" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.565112 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5df297-871d-4f7b-ac74-616325318f26" containerName="nova-metadata-log" Dec 04 04:17:23 crc kubenswrapper[4806]: E1204 04:17:23.565135 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5df297-871d-4f7b-ac74-616325318f26" containerName="nova-metadata-metadata" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.565143 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5df297-871d-4f7b-ac74-616325318f26" containerName="nova-metadata-metadata" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.565419 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce5df297-871d-4f7b-ac74-616325318f26" containerName="nova-metadata-log" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.565445 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec8caa98-4401-4f0b-86fa-ff3d04e00544" containerName="nova-cell1-conductor-db-sync" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.565456 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce5df297-871d-4f7b-ac74-616325318f26" containerName="nova-metadata-metadata" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.566259 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.569571 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.586546 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.604728 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.677997 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.718433 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.720471 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.725000 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.725606 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.753104 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.770233 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7137b5c-7316-46a0-abda-976ee89ac4ac-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e7137b5c-7316-46a0-abda-976ee89ac4ac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.770371 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7137b5c-7316-46a0-abda-976ee89ac4ac-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e7137b5c-7316-46a0-abda-976ee89ac4ac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.770426 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smqn2\" (UniqueName: \"kubernetes.io/projected/e7137b5c-7316-46a0-abda-976ee89ac4ac-kube-api-access-smqn2\") pod \"nova-cell1-conductor-0\" (UID: \"e7137b5c-7316-46a0-abda-976ee89ac4ac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 04:17:23 crc kubenswrapper[4806]: E1204 04:17:23.821376 4806 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec8caa98_4401_4f0b_86fa_ff3d04e00544.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce5df297_871d_4f7b_ac74_616325318f26.slice/crio-2604b886d8a6b8b11be8477266980fe1aa0a9ea6cff59e8d327aaec4d8afe894\": RecentStats: unable to find data in memory cache]" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.873717 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smqn2\" (UniqueName: \"kubernetes.io/projected/e7137b5c-7316-46a0-abda-976ee89ac4ac-kube-api-access-smqn2\") pod \"nova-cell1-conductor-0\" (UID: \"e7137b5c-7316-46a0-abda-976ee89ac4ac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.874626 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-config-data\") pod \"nova-metadata-0\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " pod="openstack/nova-metadata-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.874794 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " pod="openstack/nova-metadata-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.874842 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fwrs\" (UniqueName: \"kubernetes.io/projected/417de6e8-85e9-47d5-843a-c44096913ccb-kube-api-access-8fwrs\") pod \"nova-metadata-0\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " pod="openstack/nova-metadata-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.874883 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7137b5c-7316-46a0-abda-976ee89ac4ac-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e7137b5c-7316-46a0-abda-976ee89ac4ac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.876023 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/417de6e8-85e9-47d5-843a-c44096913ccb-logs\") pod \"nova-metadata-0\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " pod="openstack/nova-metadata-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.876168 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7137b5c-7316-46a0-abda-976ee89ac4ac-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e7137b5c-7316-46a0-abda-976ee89ac4ac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.876222 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " pod="openstack/nova-metadata-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.881886 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7137b5c-7316-46a0-abda-976ee89ac4ac-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"e7137b5c-7316-46a0-abda-976ee89ac4ac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.892209 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7137b5c-7316-46a0-abda-976ee89ac4ac-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"e7137b5c-7316-46a0-abda-976ee89ac4ac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.900647 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smqn2\" (UniqueName: \"kubernetes.io/projected/e7137b5c-7316-46a0-abda-976ee89ac4ac-kube-api-access-smqn2\") pod \"nova-cell1-conductor-0\" (UID: \"e7137b5c-7316-46a0-abda-976ee89ac4ac\") " pod="openstack/nova-cell1-conductor-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.923775 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.924479 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerName="proxy-httpd" containerID="cri-o://a3de0ca96e3c93a15d66d81d98dc2fa8e933d7882004763a2cfcbd34e40d097d" gracePeriod=30 Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.924728 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerName="ceilometer-central-agent" containerID="cri-o://debff5ed08eae6bbd324d0ede5392068b6032a17a9abe84c3d1ddec2d5007110" gracePeriod=30 Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.924863 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerName="sg-core" containerID="cri-o://e4c78240de1bbe2e3316fc9be8a36d8fb4ca734ae4b31c986e12c0f85331c808" gracePeriod=30 Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.924914 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerName="ceilometer-notification-agent" containerID="cri-o://06cc56a3b6332c9c218a295b5a11d8b0037b70ee8c4bf7477fe5708cfe28094a" gracePeriod=30 Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.932568 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.978189 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-config-data\") pod \"nova-metadata-0\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " pod="openstack/nova-metadata-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.978324 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " pod="openstack/nova-metadata-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.978357 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fwrs\" (UniqueName: \"kubernetes.io/projected/417de6e8-85e9-47d5-843a-c44096913ccb-kube-api-access-8fwrs\") pod \"nova-metadata-0\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " pod="openstack/nova-metadata-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.978403 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/417de6e8-85e9-47d5-843a-c44096913ccb-logs\") pod \"nova-metadata-0\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " pod="openstack/nova-metadata-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.978443 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " pod="openstack/nova-metadata-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.979505 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/417de6e8-85e9-47d5-843a-c44096913ccb-logs\") pod \"nova-metadata-0\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " pod="openstack/nova-metadata-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.990241 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " pod="openstack/nova-metadata-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.990604 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-config-data\") pod \"nova-metadata-0\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " pod="openstack/nova-metadata-0" Dec 04 04:17:23 crc kubenswrapper[4806]: I1204 04:17:23.994204 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " pod="openstack/nova-metadata-0" Dec 04 04:17:24 crc kubenswrapper[4806]: I1204 04:17:24.006507 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fwrs\" (UniqueName: \"kubernetes.io/projected/417de6e8-85e9-47d5-843a-c44096913ccb-kube-api-access-8fwrs\") pod \"nova-metadata-0\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " pod="openstack/nova-metadata-0" Dec 04 04:17:24 crc kubenswrapper[4806]: I1204 04:17:24.048764 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 04:17:24 crc kubenswrapper[4806]: I1204 04:17:24.452323 4806 generic.go:334] "Generic (PLEG): container finished" podID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerID="a3de0ca96e3c93a15d66d81d98dc2fa8e933d7882004763a2cfcbd34e40d097d" exitCode=0 Dec 04 04:17:24 crc kubenswrapper[4806]: I1204 04:17:24.452683 4806 generic.go:334] "Generic (PLEG): container finished" podID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerID="e4c78240de1bbe2e3316fc9be8a36d8fb4ca734ae4b31c986e12c0f85331c808" exitCode=2 Dec 04 04:17:24 crc kubenswrapper[4806]: I1204 04:17:24.453467 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18170ea9-c07e-4bf4-b6d4-5f72317b3e95","Type":"ContainerDied","Data":"a3de0ca96e3c93a15d66d81d98dc2fa8e933d7882004763a2cfcbd34e40d097d"} Dec 04 04:17:24 crc kubenswrapper[4806]: I1204 04:17:24.453510 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18170ea9-c07e-4bf4-b6d4-5f72317b3e95","Type":"ContainerDied","Data":"e4c78240de1bbe2e3316fc9be8a36d8fb4ca734ae4b31c986e12c0f85331c808"} Dec 04 04:17:24 crc kubenswrapper[4806]: I1204 04:17:24.557526 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 04:17:24 crc kubenswrapper[4806]: I1204 04:17:24.698700 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.440417 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce5df297-871d-4f7b-ac74-616325318f26" path="/var/lib/kubelet/pods/ce5df297-871d-4f7b-ac74-616325318f26/volumes" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.478898 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e7137b5c-7316-46a0-abda-976ee89ac4ac","Type":"ContainerStarted","Data":"ba59fe1cd5d5c472d9c9dae8c85c7f676ac4d5aba4b8f83f91b7256d7c025e70"} Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.478972 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"e7137b5c-7316-46a0-abda-976ee89ac4ac","Type":"ContainerStarted","Data":"671da535e7ba68c27f3520070ba1329cfbb0132cb779d07d94eb14e271f2f471"} Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.480202 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.483218 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.484744 4806 generic.go:334] "Generic (PLEG): container finished" podID="f435c60b-81d5-41a0-870a-d2825cc97779" containerID="3f8c897cb5bfdc9b31b97b8f517383483eb63cf329c035bd24e4c97d97c61359" exitCode=0 Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.484845 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f435c60b-81d5-41a0-870a-d2825cc97779","Type":"ContainerDied","Data":"3f8c897cb5bfdc9b31b97b8f517383483eb63cf329c035bd24e4c97d97c61359"} Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.484889 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f435c60b-81d5-41a0-870a-d2825cc97779","Type":"ContainerDied","Data":"7601b85f1c8dd979d689399aacb8e11ed7320bbe2a7e0244b393aeda9eef73bc"} Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.484914 4806 scope.go:117] "RemoveContainer" containerID="3f8c897cb5bfdc9b31b97b8f517383483eb63cf329c035bd24e4c97d97c61359" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.493612 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"417de6e8-85e9-47d5-843a-c44096913ccb","Type":"ContainerStarted","Data":"fdd0d0aa5b3c5b4c7be455ca09c59eada2a19044b1bae433a1366338176d2220"} Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.493681 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"417de6e8-85e9-47d5-843a-c44096913ccb","Type":"ContainerStarted","Data":"d1874375f111276dce66e7b66c976108a8db4ebfd21470a75921ac1b2011db2e"} Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.493696 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"417de6e8-85e9-47d5-843a-c44096913ccb","Type":"ContainerStarted","Data":"9d700145d0bb43156fc9df127784d31545e71cbfd9e588b9b03f34a79b7b305b"} Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.522517 4806 generic.go:334] "Generic (PLEG): container finished" podID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerID="debff5ed08eae6bbd324d0ede5392068b6032a17a9abe84c3d1ddec2d5007110" exitCode=0 Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.522575 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18170ea9-c07e-4bf4-b6d4-5f72317b3e95","Type":"ContainerDied","Data":"debff5ed08eae6bbd324d0ede5392068b6032a17a9abe84c3d1ddec2d5007110"} Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.534853 4806 scope.go:117] "RemoveContainer" containerID="e94a9577e68c4802411798687bcaad71b8a6911f920c85a685ad6d906c016e77" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.545681 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.545660362 podStartE2EDuration="2.545660362s" podCreationTimestamp="2025-12-04 04:17:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:17:25.523245114 +0000 UTC m=+1360.381758032" watchObservedRunningTime="2025-12-04 04:17:25.545660362 +0000 UTC m=+1360.404173290" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.571574 4806 scope.go:117] "RemoveContainer" containerID="3f8c897cb5bfdc9b31b97b8f517383483eb63cf329c035bd24e4c97d97c61359" Dec 04 04:17:25 crc kubenswrapper[4806]: E1204 04:17:25.575006 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f8c897cb5bfdc9b31b97b8f517383483eb63cf329c035bd24e4c97d97c61359\": container with ID starting with 3f8c897cb5bfdc9b31b97b8f517383483eb63cf329c035bd24e4c97d97c61359 not found: ID does not exist" containerID="3f8c897cb5bfdc9b31b97b8f517383483eb63cf329c035bd24e4c97d97c61359" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.575039 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f8c897cb5bfdc9b31b97b8f517383483eb63cf329c035bd24e4c97d97c61359"} err="failed to get container status \"3f8c897cb5bfdc9b31b97b8f517383483eb63cf329c035bd24e4c97d97c61359\": rpc error: code = NotFound desc = could not find container \"3f8c897cb5bfdc9b31b97b8f517383483eb63cf329c035bd24e4c97d97c61359\": container with ID starting with 3f8c897cb5bfdc9b31b97b8f517383483eb63cf329c035bd24e4c97d97c61359 not found: ID does not exist" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.575062 4806 scope.go:117] "RemoveContainer" containerID="e94a9577e68c4802411798687bcaad71b8a6911f920c85a685ad6d906c016e77" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.578029 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.578005912 podStartE2EDuration="2.578005912s" podCreationTimestamp="2025-12-04 04:17:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:17:25.571517817 +0000 UTC m=+1360.430030745" watchObservedRunningTime="2025-12-04 04:17:25.578005912 +0000 UTC m=+1360.436518840" Dec 04 04:17:25 crc kubenswrapper[4806]: E1204 04:17:25.581514 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e94a9577e68c4802411798687bcaad71b8a6911f920c85a685ad6d906c016e77\": container with ID starting with e94a9577e68c4802411798687bcaad71b8a6911f920c85a685ad6d906c016e77 not found: ID does not exist" containerID="e94a9577e68c4802411798687bcaad71b8a6911f920c85a685ad6d906c016e77" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.581574 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e94a9577e68c4802411798687bcaad71b8a6911f920c85a685ad6d906c016e77"} err="failed to get container status \"e94a9577e68c4802411798687bcaad71b8a6911f920c85a685ad6d906c016e77\": rpc error: code = NotFound desc = could not find container \"e94a9577e68c4802411798687bcaad71b8a6911f920c85a685ad6d906c016e77\": container with ID starting with e94a9577e68c4802411798687bcaad71b8a6911f920c85a685ad6d906c016e77 not found: ID does not exist" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.645827 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f435c60b-81d5-41a0-870a-d2825cc97779-config-data\") pod \"f435c60b-81d5-41a0-870a-d2825cc97779\" (UID: \"f435c60b-81d5-41a0-870a-d2825cc97779\") " Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.645896 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2mn5\" (UniqueName: \"kubernetes.io/projected/f435c60b-81d5-41a0-870a-d2825cc97779-kube-api-access-p2mn5\") pod \"f435c60b-81d5-41a0-870a-d2825cc97779\" (UID: \"f435c60b-81d5-41a0-870a-d2825cc97779\") " Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.646139 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f435c60b-81d5-41a0-870a-d2825cc97779-combined-ca-bundle\") pod \"f435c60b-81d5-41a0-870a-d2825cc97779\" (UID: \"f435c60b-81d5-41a0-870a-d2825cc97779\") " Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.646167 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f435c60b-81d5-41a0-870a-d2825cc97779-logs\") pod \"f435c60b-81d5-41a0-870a-d2825cc97779\" (UID: \"f435c60b-81d5-41a0-870a-d2825cc97779\") " Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.647363 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f435c60b-81d5-41a0-870a-d2825cc97779-logs" (OuterVolumeSpecName: "logs") pod "f435c60b-81d5-41a0-870a-d2825cc97779" (UID: "f435c60b-81d5-41a0-870a-d2825cc97779"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.655748 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f435c60b-81d5-41a0-870a-d2825cc97779-kube-api-access-p2mn5" (OuterVolumeSpecName: "kube-api-access-p2mn5") pod "f435c60b-81d5-41a0-870a-d2825cc97779" (UID: "f435c60b-81d5-41a0-870a-d2825cc97779"). InnerVolumeSpecName "kube-api-access-p2mn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.701180 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f435c60b-81d5-41a0-870a-d2825cc97779-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f435c60b-81d5-41a0-870a-d2825cc97779" (UID: "f435c60b-81d5-41a0-870a-d2825cc97779"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.705525 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f435c60b-81d5-41a0-870a-d2825cc97779-config-data" (OuterVolumeSpecName: "config-data") pod "f435c60b-81d5-41a0-870a-d2825cc97779" (UID: "f435c60b-81d5-41a0-870a-d2825cc97779"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.748960 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f435c60b-81d5-41a0-870a-d2825cc97779-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.749007 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f435c60b-81d5-41a0-870a-d2825cc97779-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.749024 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f435c60b-81d5-41a0-870a-d2825cc97779-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:25 crc kubenswrapper[4806]: I1204 04:17:25.749037 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2mn5\" (UniqueName: \"kubernetes.io/projected/f435c60b-81d5-41a0-870a-d2825cc97779-kube-api-access-p2mn5\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.540652 4806 generic.go:334] "Generic (PLEG): container finished" podID="6c7c39b0-17d1-4f72-993e-98192c710d83" containerID="d08d6e9a7b811235a1a7c85c2244f509e4a26b25d21c0b4b707d4d6e3bb75dbf" exitCode=0 Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.540986 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6c7c39b0-17d1-4f72-993e-98192c710d83","Type":"ContainerDied","Data":"d08d6e9a7b811235a1a7c85c2244f509e4a26b25d21c0b4b707d4d6e3bb75dbf"} Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.545291 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.609989 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.627427 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:17:26 crc kubenswrapper[4806]: E1204 04:17:26.637462 4806 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d08d6e9a7b811235a1a7c85c2244f509e4a26b25d21c0b4b707d4d6e3bb75dbf is running failed: container process not found" containerID="d08d6e9a7b811235a1a7c85c2244f509e4a26b25d21c0b4b707d4d6e3bb75dbf" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 04:17:26 crc kubenswrapper[4806]: E1204 04:17:26.641081 4806 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d08d6e9a7b811235a1a7c85c2244f509e4a26b25d21c0b4b707d4d6e3bb75dbf is running failed: container process not found" containerID="d08d6e9a7b811235a1a7c85c2244f509e4a26b25d21c0b4b707d4d6e3bb75dbf" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.641273 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 04:17:26 crc kubenswrapper[4806]: E1204 04:17:26.641751 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f435c60b-81d5-41a0-870a-d2825cc97779" containerName="nova-api-log" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.641770 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f435c60b-81d5-41a0-870a-d2825cc97779" containerName="nova-api-log" Dec 04 04:17:26 crc kubenswrapper[4806]: E1204 04:17:26.641784 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f435c60b-81d5-41a0-870a-d2825cc97779" containerName="nova-api-api" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.641791 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f435c60b-81d5-41a0-870a-d2825cc97779" containerName="nova-api-api" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.641986 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f435c60b-81d5-41a0-870a-d2825cc97779" containerName="nova-api-log" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.642007 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f435c60b-81d5-41a0-870a-d2825cc97779" containerName="nova-api-api" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.643027 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 04:17:26 crc kubenswrapper[4806]: E1204 04:17:26.645722 4806 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d08d6e9a7b811235a1a7c85c2244f509e4a26b25d21c0b4b707d4d6e3bb75dbf is running failed: container process not found" containerID="d08d6e9a7b811235a1a7c85c2244f509e4a26b25d21c0b4b707d4d6e3bb75dbf" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.648403 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 04:17:26 crc kubenswrapper[4806]: E1204 04:17:26.649131 4806 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d08d6e9a7b811235a1a7c85c2244f509e4a26b25d21c0b4b707d4d6e3bb75dbf is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="6c7c39b0-17d1-4f72-993e-98192c710d83" containerName="nova-scheduler-scheduler" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.655265 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.776843 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-config-data\") pod \"nova-api-0\" (UID: \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\") " pod="openstack/nova-api-0" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.777123 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ztqp\" (UniqueName: \"kubernetes.io/projected/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-kube-api-access-4ztqp\") pod \"nova-api-0\" (UID: \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\") " pod="openstack/nova-api-0" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.777501 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-logs\") pod \"nova-api-0\" (UID: \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\") " pod="openstack/nova-api-0" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.777723 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\") " pod="openstack/nova-api-0" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.880049 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\") " pod="openstack/nova-api-0" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.880782 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-config-data\") pod \"nova-api-0\" (UID: \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\") " pod="openstack/nova-api-0" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.880846 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ztqp\" (UniqueName: \"kubernetes.io/projected/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-kube-api-access-4ztqp\") pod \"nova-api-0\" (UID: \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\") " pod="openstack/nova-api-0" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.881647 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-logs\") pod \"nova-api-0\" (UID: \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\") " pod="openstack/nova-api-0" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.882151 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-logs\") pod \"nova-api-0\" (UID: \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\") " pod="openstack/nova-api-0" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.899687 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-config-data\") pod \"nova-api-0\" (UID: \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\") " pod="openstack/nova-api-0" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.913912 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ztqp\" (UniqueName: \"kubernetes.io/projected/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-kube-api-access-4ztqp\") pod \"nova-api-0\" (UID: \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\") " pod="openstack/nova-api-0" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.918057 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\") " pod="openstack/nova-api-0" Dec 04 04:17:26 crc kubenswrapper[4806]: I1204 04:17:26.974236 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.113470 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.287655 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjb7q\" (UniqueName: \"kubernetes.io/projected/6c7c39b0-17d1-4f72-993e-98192c710d83-kube-api-access-jjb7q\") pod \"6c7c39b0-17d1-4f72-993e-98192c710d83\" (UID: \"6c7c39b0-17d1-4f72-993e-98192c710d83\") " Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.287778 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c7c39b0-17d1-4f72-993e-98192c710d83-config-data\") pod \"6c7c39b0-17d1-4f72-993e-98192c710d83\" (UID: \"6c7c39b0-17d1-4f72-993e-98192c710d83\") " Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.287878 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c7c39b0-17d1-4f72-993e-98192c710d83-combined-ca-bundle\") pod \"6c7c39b0-17d1-4f72-993e-98192c710d83\" (UID: \"6c7c39b0-17d1-4f72-993e-98192c710d83\") " Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.317658 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c7c39b0-17d1-4f72-993e-98192c710d83-kube-api-access-jjb7q" (OuterVolumeSpecName: "kube-api-access-jjb7q") pod "6c7c39b0-17d1-4f72-993e-98192c710d83" (UID: "6c7c39b0-17d1-4f72-993e-98192c710d83"). InnerVolumeSpecName "kube-api-access-jjb7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.345305 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c7c39b0-17d1-4f72-993e-98192c710d83-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c7c39b0-17d1-4f72-993e-98192c710d83" (UID: "6c7c39b0-17d1-4f72-993e-98192c710d83"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.377054 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c7c39b0-17d1-4f72-993e-98192c710d83-config-data" (OuterVolumeSpecName: "config-data") pod "6c7c39b0-17d1-4f72-993e-98192c710d83" (UID: "6c7c39b0-17d1-4f72-993e-98192c710d83"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.390530 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjb7q\" (UniqueName: \"kubernetes.io/projected/6c7c39b0-17d1-4f72-993e-98192c710d83-kube-api-access-jjb7q\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.390565 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c7c39b0-17d1-4f72-993e-98192c710d83-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.390577 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c7c39b0-17d1-4f72-993e-98192c710d83-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.434589 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f435c60b-81d5-41a0-870a-d2825cc97779" path="/var/lib/kubelet/pods/f435c60b-81d5-41a0-870a-d2825cc97779/volumes" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.496564 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.562128 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6c7c39b0-17d1-4f72-993e-98192c710d83","Type":"ContainerDied","Data":"c4498b068e9da95c8df36bf4775a1a05096d9ace4f0a8c62f50f302602876c1f"} Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.562187 4806 scope.go:117] "RemoveContainer" containerID="d08d6e9a7b811235a1a7c85c2244f509e4a26b25d21c0b4b707d4d6e3bb75dbf" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.562139 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.566605 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1354ec28-6b19-4f91-9b52-fbcfd83b6fce","Type":"ContainerStarted","Data":"53efd9e51bb5305bdf228205a677ed461f457379fe877f752b84c3b820d8ff3a"} Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.668157 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.685645 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.693030 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 04:17:27 crc kubenswrapper[4806]: E1204 04:17:27.693619 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c7c39b0-17d1-4f72-993e-98192c710d83" containerName="nova-scheduler-scheduler" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.693652 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c7c39b0-17d1-4f72-993e-98192c710d83" containerName="nova-scheduler-scheduler" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.693866 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c7c39b0-17d1-4f72-993e-98192c710d83" containerName="nova-scheduler-scheduler" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.694886 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.699053 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.703766 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.802323 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77484080-5dac-49d3-b5c7-96aefe955966-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"77484080-5dac-49d3-b5c7-96aefe955966\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.802387 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwqhx\" (UniqueName: \"kubernetes.io/projected/77484080-5dac-49d3-b5c7-96aefe955966-kube-api-access-gwqhx\") pod \"nova-scheduler-0\" (UID: \"77484080-5dac-49d3-b5c7-96aefe955966\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.802447 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77484080-5dac-49d3-b5c7-96aefe955966-config-data\") pod \"nova-scheduler-0\" (UID: \"77484080-5dac-49d3-b5c7-96aefe955966\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.904560 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77484080-5dac-49d3-b5c7-96aefe955966-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"77484080-5dac-49d3-b5c7-96aefe955966\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.904982 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwqhx\" (UniqueName: \"kubernetes.io/projected/77484080-5dac-49d3-b5c7-96aefe955966-kube-api-access-gwqhx\") pod \"nova-scheduler-0\" (UID: \"77484080-5dac-49d3-b5c7-96aefe955966\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.905055 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77484080-5dac-49d3-b5c7-96aefe955966-config-data\") pod \"nova-scheduler-0\" (UID: \"77484080-5dac-49d3-b5c7-96aefe955966\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.909069 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77484080-5dac-49d3-b5c7-96aefe955966-config-data\") pod \"nova-scheduler-0\" (UID: \"77484080-5dac-49d3-b5c7-96aefe955966\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.911653 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77484080-5dac-49d3-b5c7-96aefe955966-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"77484080-5dac-49d3-b5c7-96aefe955966\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:27 crc kubenswrapper[4806]: I1204 04:17:27.928774 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwqhx\" (UniqueName: \"kubernetes.io/projected/77484080-5dac-49d3-b5c7-96aefe955966-kube-api-access-gwqhx\") pod \"nova-scheduler-0\" (UID: \"77484080-5dac-49d3-b5c7-96aefe955966\") " pod="openstack/nova-scheduler-0" Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.096047 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.587967 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1354ec28-6b19-4f91-9b52-fbcfd83b6fce","Type":"ContainerStarted","Data":"df2bcdcd0af1a361583a7328f1abb4da9b4bda4bc14ec666e73c34e16f5e6dc2"} Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.588360 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1354ec28-6b19-4f91-9b52-fbcfd83b6fce","Type":"ContainerStarted","Data":"73b3aa0ec02f6571b04650e56fb636fc2ccacc3fc01c56d6299de8613ca357ca"} Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.626935 4806 generic.go:334] "Generic (PLEG): container finished" podID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerID="06cc56a3b6332c9c218a295b5a11d8b0037b70ee8c4bf7477fe5708cfe28094a" exitCode=0 Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.627002 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18170ea9-c07e-4bf4-b6d4-5f72317b3e95","Type":"ContainerDied","Data":"06cc56a3b6332c9c218a295b5a11d8b0037b70ee8c4bf7477fe5708cfe28094a"} Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.639152 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.639135337 podStartE2EDuration="2.639135337s" podCreationTimestamp="2025-12-04 04:17:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:17:28.613070241 +0000 UTC m=+1363.471583169" watchObservedRunningTime="2025-12-04 04:17:28.639135337 +0000 UTC m=+1363.497648255" Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.640765 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.811123 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.947273 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-combined-ca-bundle\") pod \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.947601 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-sg-core-conf-yaml\") pod \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.947650 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-config-data\") pod \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.947681 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-run-httpd\") pod \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.947761 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9tqk\" (UniqueName: \"kubernetes.io/projected/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-kube-api-access-s9tqk\") pod \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.947796 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-log-httpd\") pod \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.947904 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-scripts\") pod \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\" (UID: \"18170ea9-c07e-4bf4-b6d4-5f72317b3e95\") " Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.949987 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "18170ea9-c07e-4bf4-b6d4-5f72317b3e95" (UID: "18170ea9-c07e-4bf4-b6d4-5f72317b3e95"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.951245 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "18170ea9-c07e-4bf4-b6d4-5f72317b3e95" (UID: "18170ea9-c07e-4bf4-b6d4-5f72317b3e95"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.958128 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-kube-api-access-s9tqk" (OuterVolumeSpecName: "kube-api-access-s9tqk") pod "18170ea9-c07e-4bf4-b6d4-5f72317b3e95" (UID: "18170ea9-c07e-4bf4-b6d4-5f72317b3e95"). InnerVolumeSpecName "kube-api-access-s9tqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:17:28 crc kubenswrapper[4806]: I1204 04:17:28.959065 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-scripts" (OuterVolumeSpecName: "scripts") pod "18170ea9-c07e-4bf4-b6d4-5f72317b3e95" (UID: "18170ea9-c07e-4bf4-b6d4-5f72317b3e95"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.017898 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "18170ea9-c07e-4bf4-b6d4-5f72317b3e95" (UID: "18170ea9-c07e-4bf4-b6d4-5f72317b3e95"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.056047 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.057171 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.057904 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.057960 4806 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.057983 4806 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.057992 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9tqk\" (UniqueName: \"kubernetes.io/projected/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-kube-api-access-s9tqk\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.058001 4806 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.069270 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18170ea9-c07e-4bf4-b6d4-5f72317b3e95" (UID: "18170ea9-c07e-4bf4-b6d4-5f72317b3e95"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.141101 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-config-data" (OuterVolumeSpecName: "config-data") pod "18170ea9-c07e-4bf4-b6d4-5f72317b3e95" (UID: "18170ea9-c07e-4bf4-b6d4-5f72317b3e95"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.159738 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.159796 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18170ea9-c07e-4bf4-b6d4-5f72317b3e95-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.434157 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c7c39b0-17d1-4f72-993e-98192c710d83" path="/var/lib/kubelet/pods/6c7c39b0-17d1-4f72-993e-98192c710d83/volumes" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.652782 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18170ea9-c07e-4bf4-b6d4-5f72317b3e95","Type":"ContainerDied","Data":"f78e6b536689f2b28dc668a28cf78ec4ac5a5c7054fd36eb792db539b6ed8365"} Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.652828 4806 scope.go:117] "RemoveContainer" containerID="a3de0ca96e3c93a15d66d81d98dc2fa8e933d7882004763a2cfcbd34e40d097d" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.652950 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.658244 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"77484080-5dac-49d3-b5c7-96aefe955966","Type":"ContainerStarted","Data":"8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6"} Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.658295 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"77484080-5dac-49d3-b5c7-96aefe955966","Type":"ContainerStarted","Data":"2bf2215d7fa6e06f3699823476db2070b86d4b36a6a0573e474781381c171a16"} Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.694242 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.706940 4806 scope.go:117] "RemoveContainer" containerID="e4c78240de1bbe2e3316fc9be8a36d8fb4ca734ae4b31c986e12c0f85331c808" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.720274 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.736465 4806 scope.go:117] "RemoveContainer" containerID="06cc56a3b6332c9c218a295b5a11d8b0037b70ee8c4bf7477fe5708cfe28094a" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.741798 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:17:29 crc kubenswrapper[4806]: E1204 04:17:29.742234 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerName="sg-core" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.742245 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerName="sg-core" Dec 04 04:17:29 crc kubenswrapper[4806]: E1204 04:17:29.742272 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerName="ceilometer-central-agent" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.742278 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerName="ceilometer-central-agent" Dec 04 04:17:29 crc kubenswrapper[4806]: E1204 04:17:29.742297 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerName="proxy-httpd" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.742303 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerName="proxy-httpd" Dec 04 04:17:29 crc kubenswrapper[4806]: E1204 04:17:29.742317 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerName="ceilometer-notification-agent" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.742322 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerName="ceilometer-notification-agent" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.742528 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerName="ceilometer-central-agent" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.742551 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerName="ceilometer-notification-agent" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.742572 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerName="sg-core" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.742582 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" containerName="proxy-httpd" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.744441 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.745353 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.745334819 podStartE2EDuration="2.745334819s" podCreationTimestamp="2025-12-04 04:17:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:17:29.701426419 +0000 UTC m=+1364.559939347" watchObservedRunningTime="2025-12-04 04:17:29.745334819 +0000 UTC m=+1364.603847747" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.748189 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.761163 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.761424 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.775316 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.798122 4806 scope.go:117] "RemoveContainer" containerID="debff5ed08eae6bbd324d0ede5392068b6032a17a9abe84c3d1ddec2d5007110" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.885536 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-config-data\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.885610 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/528ecdc7-e743-41de-b060-5aec0be99851-run-httpd\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.885654 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.885893 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-scripts\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.886031 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.886084 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.886345 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqfsw\" (UniqueName: \"kubernetes.io/projected/528ecdc7-e743-41de-b060-5aec0be99851-kube-api-access-qqfsw\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.886401 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/528ecdc7-e743-41de-b060-5aec0be99851-log-httpd\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.949443 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bc4f74d8d-xj2xn" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.987874 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqfsw\" (UniqueName: \"kubernetes.io/projected/528ecdc7-e743-41de-b060-5aec0be99851-kube-api-access-qqfsw\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.987945 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/528ecdc7-e743-41de-b060-5aec0be99851-log-httpd\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.988007 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-config-data\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.988038 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/528ecdc7-e743-41de-b060-5aec0be99851-run-httpd\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.988061 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.988650 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/528ecdc7-e743-41de-b060-5aec0be99851-log-httpd\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.988678 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/528ecdc7-e743-41de-b060-5aec0be99851-run-httpd\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.988128 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-scripts\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.989037 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.989066 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.997589 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-scripts\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:29 crc kubenswrapper[4806]: I1204 04:17:29.999202 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-config-data\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:30 crc kubenswrapper[4806]: I1204 04:17:30.003519 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:30 crc kubenswrapper[4806]: I1204 04:17:30.005460 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:30 crc kubenswrapper[4806]: I1204 04:17:30.009504 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:30 crc kubenswrapper[4806]: I1204 04:17:30.014458 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqfsw\" (UniqueName: \"kubernetes.io/projected/528ecdc7-e743-41de-b060-5aec0be99851-kube-api-access-qqfsw\") pod \"ceilometer-0\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " pod="openstack/ceilometer-0" Dec 04 04:17:30 crc kubenswrapper[4806]: I1204 04:17:30.090718 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:17:30 crc kubenswrapper[4806]: I1204 04:17:30.379379 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d8bd6c96d-4sqhf" podUID="d2d53864-c3c8-4e40-8791-0bac49ba561e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 04:17:30 crc kubenswrapper[4806]: I1204 04:17:30.605810 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:17:30 crc kubenswrapper[4806]: W1204 04:17:30.610473 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod528ecdc7_e743_41de_b060_5aec0be99851.slice/crio-cff92425fe243adcc23e9b96d02fac4a75c4f3e2e5b03b9e3675d15758072684 WatchSource:0}: Error finding container cff92425fe243adcc23e9b96d02fac4a75c4f3e2e5b03b9e3675d15758072684: Status 404 returned error can't find the container with id cff92425fe243adcc23e9b96d02fac4a75c4f3e2e5b03b9e3675d15758072684 Dec 04 04:17:30 crc kubenswrapper[4806]: I1204 04:17:30.669076 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"528ecdc7-e743-41de-b060-5aec0be99851","Type":"ContainerStarted","Data":"cff92425fe243adcc23e9b96d02fac4a75c4f3e2e5b03b9e3675d15758072684"} Dec 04 04:17:31 crc kubenswrapper[4806]: I1204 04:17:31.439909 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18170ea9-c07e-4bf4-b6d4-5f72317b3e95" path="/var/lib/kubelet/pods/18170ea9-c07e-4bf4-b6d4-5f72317b3e95/volumes" Dec 04 04:17:31 crc kubenswrapper[4806]: I1204 04:17:31.680469 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"528ecdc7-e743-41de-b060-5aec0be99851","Type":"ContainerStarted","Data":"34d312119bbdf650410a1fadafe3de601e99be71ecd1ee5371fb4a820e14d05f"} Dec 04 04:17:31 crc kubenswrapper[4806]: I1204 04:17:31.791497 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 04 04:17:32 crc kubenswrapper[4806]: I1204 04:17:32.694572 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"528ecdc7-e743-41de-b060-5aec0be99851","Type":"ContainerStarted","Data":"84e16224ae48b3f3b23576debccd8b560aa3741a55ce07cc2d3079aac5f187a7"} Dec 04 04:17:32 crc kubenswrapper[4806]: I1204 04:17:32.695363 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"528ecdc7-e743-41de-b060-5aec0be99851","Type":"ContainerStarted","Data":"d0e8cb77705a9d74725495026ec255cb2b9da5d4bbcb6afcec32fd945fc118b9"} Dec 04 04:17:33 crc kubenswrapper[4806]: I1204 04:17:33.096226 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 04:17:33 crc kubenswrapper[4806]: I1204 04:17:33.986188 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 04 04:17:34 crc kubenswrapper[4806]: I1204 04:17:34.052011 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 04:17:34 crc kubenswrapper[4806]: I1204 04:17:34.052284 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 04:17:34 crc kubenswrapper[4806]: I1204 04:17:34.718688 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"528ecdc7-e743-41de-b060-5aec0be99851","Type":"ContainerStarted","Data":"8aa53bcf517e462fd9a64d48de52fd9484d3a2861af550e380a93ce6b9d97ec1"} Dec 04 04:17:34 crc kubenswrapper[4806]: I1204 04:17:34.719091 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 04:17:34 crc kubenswrapper[4806]: I1204 04:17:34.740181 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.324666168 podStartE2EDuration="5.740162564s" podCreationTimestamp="2025-12-04 04:17:29 +0000 UTC" firstStartedPulling="2025-12-04 04:17:30.612221678 +0000 UTC m=+1365.470734616" lastFinishedPulling="2025-12-04 04:17:34.027718084 +0000 UTC m=+1368.886231012" observedRunningTime="2025-12-04 04:17:34.737260162 +0000 UTC m=+1369.595773090" watchObservedRunningTime="2025-12-04 04:17:34.740162564 +0000 UTC m=+1369.598675492" Dec 04 04:17:35 crc kubenswrapper[4806]: I1204 04:17:35.061281 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="417de6e8-85e9-47d5-843a-c44096913ccb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:17:35 crc kubenswrapper[4806]: I1204 04:17:35.070123 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="417de6e8-85e9-47d5-843a-c44096913ccb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:17:36 crc kubenswrapper[4806]: I1204 04:17:36.975384 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 04:17:36 crc kubenswrapper[4806]: I1204 04:17:36.975662 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 04:17:38 crc kubenswrapper[4806]: I1204 04:17:38.059183 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1354ec28-6b19-4f91-9b52-fbcfd83b6fce" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 04:17:38 crc kubenswrapper[4806]: I1204 04:17:38.059197 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1354ec28-6b19-4f91-9b52-fbcfd83b6fce" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 04:17:38 crc kubenswrapper[4806]: I1204 04:17:38.097343 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 04:17:38 crc kubenswrapper[4806]: I1204 04:17:38.133099 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 04:17:38 crc kubenswrapper[4806]: I1204 04:17:38.788569 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 04:17:39 crc kubenswrapper[4806]: I1204 04:17:39.943684 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bc4f74d8d-xj2xn" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 04 04:17:40 crc kubenswrapper[4806]: I1204 04:17:40.377167 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d8bd6c96d-4sqhf" podUID="d2d53864-c3c8-4e40-8791-0bac49ba561e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.567029 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.665701 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-combined-ca-bundle\") pod \"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85\" (UID: \"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85\") " Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.665900 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-config-data\") pod \"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85\" (UID: \"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85\") " Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.666089 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nhlt\" (UniqueName: \"kubernetes.io/projected/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-kube-api-access-6nhlt\") pod \"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85\" (UID: \"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85\") " Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.675399 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-kube-api-access-6nhlt" (OuterVolumeSpecName: "kube-api-access-6nhlt") pod "d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85" (UID: "d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85"). InnerVolumeSpecName "kube-api-access-6nhlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.700782 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-config-data" (OuterVolumeSpecName: "config-data") pod "d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85" (UID: "d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.703412 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85" (UID: "d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.768454 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.768984 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.769052 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nhlt\" (UniqueName: \"kubernetes.io/projected/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85-kube-api-access-6nhlt\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.926218 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85" containerID="1f3327df49ec06e892dd92691f11321a97476b1e07e1a4b55af59069d2dcf357" exitCode=137 Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.926271 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85","Type":"ContainerDied","Data":"1f3327df49ec06e892dd92691f11321a97476b1e07e1a4b55af59069d2dcf357"} Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.926305 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85","Type":"ContainerDied","Data":"d2e692527a543c58e8c0faefb4881f1af2ef99cf62c8393270d16f51997df9c6"} Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.926298 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.926330 4806 scope.go:117] "RemoveContainer" containerID="1f3327df49ec06e892dd92691f11321a97476b1e07e1a4b55af59069d2dcf357" Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.957657 4806 scope.go:117] "RemoveContainer" containerID="1f3327df49ec06e892dd92691f11321a97476b1e07e1a4b55af59069d2dcf357" Dec 04 04:17:43 crc kubenswrapper[4806]: E1204 04:17:43.958507 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f3327df49ec06e892dd92691f11321a97476b1e07e1a4b55af59069d2dcf357\": container with ID starting with 1f3327df49ec06e892dd92691f11321a97476b1e07e1a4b55af59069d2dcf357 not found: ID does not exist" containerID="1f3327df49ec06e892dd92691f11321a97476b1e07e1a4b55af59069d2dcf357" Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.958573 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f3327df49ec06e892dd92691f11321a97476b1e07e1a4b55af59069d2dcf357"} err="failed to get container status \"1f3327df49ec06e892dd92691f11321a97476b1e07e1a4b55af59069d2dcf357\": rpc error: code = NotFound desc = could not find container \"1f3327df49ec06e892dd92691f11321a97476b1e07e1a4b55af59069d2dcf357\": container with ID starting with 1f3327df49ec06e892dd92691f11321a97476b1e07e1a4b55af59069d2dcf357 not found: ID does not exist" Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.967917 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 04:17:43 crc kubenswrapper[4806]: I1204 04:17:43.991385 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.001229 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 04:17:44 crc kubenswrapper[4806]: E1204 04:17:44.001817 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.001840 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.002143 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.002968 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.005500 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.005654 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.005798 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.012149 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.056580 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.056656 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.076856 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scb6n\" (UniqueName: \"kubernetes.io/projected/3999d3e5-7b66-4792-9fd0-996d0d97efd2-kube-api-access-scb6n\") pod \"nova-cell1-novncproxy-0\" (UID: \"3999d3e5-7b66-4792-9fd0-996d0d97efd2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.076978 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3999d3e5-7b66-4792-9fd0-996d0d97efd2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3999d3e5-7b66-4792-9fd0-996d0d97efd2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.077056 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3999d3e5-7b66-4792-9fd0-996d0d97efd2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3999d3e5-7b66-4792-9fd0-996d0d97efd2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.077174 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3999d3e5-7b66-4792-9fd0-996d0d97efd2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3999d3e5-7b66-4792-9fd0-996d0d97efd2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.077209 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3999d3e5-7b66-4792-9fd0-996d0d97efd2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3999d3e5-7b66-4792-9fd0-996d0d97efd2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.078037 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.086351 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.184462 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scb6n\" (UniqueName: \"kubernetes.io/projected/3999d3e5-7b66-4792-9fd0-996d0d97efd2-kube-api-access-scb6n\") pod \"nova-cell1-novncproxy-0\" (UID: \"3999d3e5-7b66-4792-9fd0-996d0d97efd2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.184551 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3999d3e5-7b66-4792-9fd0-996d0d97efd2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3999d3e5-7b66-4792-9fd0-996d0d97efd2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.184624 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3999d3e5-7b66-4792-9fd0-996d0d97efd2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3999d3e5-7b66-4792-9fd0-996d0d97efd2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.184744 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3999d3e5-7b66-4792-9fd0-996d0d97efd2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3999d3e5-7b66-4792-9fd0-996d0d97efd2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.184769 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3999d3e5-7b66-4792-9fd0-996d0d97efd2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3999d3e5-7b66-4792-9fd0-996d0d97efd2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.192146 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3999d3e5-7b66-4792-9fd0-996d0d97efd2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3999d3e5-7b66-4792-9fd0-996d0d97efd2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.205902 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/3999d3e5-7b66-4792-9fd0-996d0d97efd2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3999d3e5-7b66-4792-9fd0-996d0d97efd2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.212416 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/3999d3e5-7b66-4792-9fd0-996d0d97efd2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"3999d3e5-7b66-4792-9fd0-996d0d97efd2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.212984 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3999d3e5-7b66-4792-9fd0-996d0d97efd2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3999d3e5-7b66-4792-9fd0-996d0d97efd2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.217011 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scb6n\" (UniqueName: \"kubernetes.io/projected/3999d3e5-7b66-4792-9fd0-996d0d97efd2-kube-api-access-scb6n\") pod \"nova-cell1-novncproxy-0\" (UID: \"3999d3e5-7b66-4792-9fd0-996d0d97efd2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.340628 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.827018 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 04:17:44 crc kubenswrapper[4806]: I1204 04:17:44.943031 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3999d3e5-7b66-4792-9fd0-996d0d97efd2","Type":"ContainerStarted","Data":"41a59630ba016b330b622bb372f083d7f722f0d9f6d15812a339caace707bad8"} Dec 04 04:17:45 crc kubenswrapper[4806]: I1204 04:17:45.438505 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85" path="/var/lib/kubelet/pods/d0c890ce-e3d2-4a9e-b2b6-6d23b6bffc85/volumes" Dec 04 04:17:45 crc kubenswrapper[4806]: I1204 04:17:45.962249 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3999d3e5-7b66-4792-9fd0-996d0d97efd2","Type":"ContainerStarted","Data":"3aec222d74e23069aded6a9e7db6f9b82e8abb2cc428d5b8b4f3fcd5a1bb77ca"} Dec 04 04:17:45 crc kubenswrapper[4806]: I1204 04:17:45.993159 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.99314084 podStartE2EDuration="2.99314084s" podCreationTimestamp="2025-12-04 04:17:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:17:45.986528181 +0000 UTC m=+1380.845041109" watchObservedRunningTime="2025-12-04 04:17:45.99314084 +0000 UTC m=+1380.851653768" Dec 04 04:17:46 crc kubenswrapper[4806]: I1204 04:17:46.978658 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 04:17:46 crc kubenswrapper[4806]: I1204 04:17:46.979616 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 04:17:46 crc kubenswrapper[4806]: I1204 04:17:46.980608 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 04:17:46 crc kubenswrapper[4806]: I1204 04:17:46.980740 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 04:17:46 crc kubenswrapper[4806]: I1204 04:17:46.984156 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 04:17:46 crc kubenswrapper[4806]: I1204 04:17:46.987627 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.271228 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-mp95p"] Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.273146 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.308578 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-mp95p"] Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.365508 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-config\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.365790 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf4fv\" (UniqueName: \"kubernetes.io/projected/66a62f3e-3d15-4d90-b725-69a85adb938d-kube-api-access-bf4fv\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.365907 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.366154 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.366223 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.366383 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.468669 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf4fv\" (UniqueName: \"kubernetes.io/projected/66a62f3e-3d15-4d90-b725-69a85adb938d-kube-api-access-bf4fv\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.468765 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.468811 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.468839 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.468880 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.468940 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-config\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.470017 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-config\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.470402 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.470560 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.470902 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.471002 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.502801 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf4fv\" (UniqueName: \"kubernetes.io/projected/66a62f3e-3d15-4d90-b725-69a85adb938d-kube-api-access-bf4fv\") pod \"dnsmasq-dns-89c5cd4d5-mp95p\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:47 crc kubenswrapper[4806]: I1204 04:17:47.598455 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:48 crc kubenswrapper[4806]: I1204 04:17:48.169359 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-mp95p"] Dec 04 04:17:48 crc kubenswrapper[4806]: W1204 04:17:48.186994 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66a62f3e_3d15_4d90_b725_69a85adb938d.slice/crio-5d39bb6776d16b002d3b100a2d20a8026b8d899701e1c9b04b40223037f38396 WatchSource:0}: Error finding container 5d39bb6776d16b002d3b100a2d20a8026b8d899701e1c9b04b40223037f38396: Status 404 returned error can't find the container with id 5d39bb6776d16b002d3b100a2d20a8026b8d899701e1c9b04b40223037f38396 Dec 04 04:17:49 crc kubenswrapper[4806]: I1204 04:17:49.004667 4806 generic.go:334] "Generic (PLEG): container finished" podID="66a62f3e-3d15-4d90-b725-69a85adb938d" containerID="109170b5a39a76faf90a107fbdfb0b74637bb8c5547b7e55f73bd0c2a1685006" exitCode=0 Dec 04 04:17:49 crc kubenswrapper[4806]: I1204 04:17:49.005407 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" event={"ID":"66a62f3e-3d15-4d90-b725-69a85adb938d","Type":"ContainerDied","Data":"109170b5a39a76faf90a107fbdfb0b74637bb8c5547b7e55f73bd0c2a1685006"} Dec 04 04:17:49 crc kubenswrapper[4806]: I1204 04:17:49.005455 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" event={"ID":"66a62f3e-3d15-4d90-b725-69a85adb938d","Type":"ContainerStarted","Data":"5d39bb6776d16b002d3b100a2d20a8026b8d899701e1c9b04b40223037f38396"} Dec 04 04:17:49 crc kubenswrapper[4806]: I1204 04:17:49.345697 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:49 crc kubenswrapper[4806]: I1204 04:17:49.915418 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:17:50 crc kubenswrapper[4806]: I1204 04:17:50.031649 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1354ec28-6b19-4f91-9b52-fbcfd83b6fce" containerName="nova-api-log" containerID="cri-o://73b3aa0ec02f6571b04650e56fb636fc2ccacc3fc01c56d6299de8613ca357ca" gracePeriod=30 Dec 04 04:17:50 crc kubenswrapper[4806]: I1204 04:17:50.032659 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" event={"ID":"66a62f3e-3d15-4d90-b725-69a85adb938d","Type":"ContainerStarted","Data":"23fada87b780f82825f808069deed276a683d5dca7a736af5a30fac5410a372b"} Dec 04 04:17:50 crc kubenswrapper[4806]: I1204 04:17:50.032702 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:50 crc kubenswrapper[4806]: I1204 04:17:50.033009 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1354ec28-6b19-4f91-9b52-fbcfd83b6fce" containerName="nova-api-api" containerID="cri-o://df2bcdcd0af1a361583a7328f1abb4da9b4bda4bc14ec666e73c34e16f5e6dc2" gracePeriod=30 Dec 04 04:17:50 crc kubenswrapper[4806]: I1204 04:17:50.063832 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" podStartSLOduration=3.063812404 podStartE2EDuration="3.063812404s" podCreationTimestamp="2025-12-04 04:17:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:17:50.063736711 +0000 UTC m=+1384.922249639" watchObservedRunningTime="2025-12-04 04:17:50.063812404 +0000 UTC m=+1384.922325332" Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.042550 4806 generic.go:334] "Generic (PLEG): container finished" podID="1354ec28-6b19-4f91-9b52-fbcfd83b6fce" containerID="73b3aa0ec02f6571b04650e56fb636fc2ccacc3fc01c56d6299de8613ca357ca" exitCode=143 Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.042647 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1354ec28-6b19-4f91-9b52-fbcfd83b6fce","Type":"ContainerDied","Data":"73b3aa0ec02f6571b04650e56fb636fc2ccacc3fc01c56d6299de8613ca357ca"} Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.132519 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.132804 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="ceilometer-central-agent" containerID="cri-o://34d312119bbdf650410a1fadafe3de601e99be71ecd1ee5371fb4a820e14d05f" gracePeriod=30 Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.133061 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="proxy-httpd" containerID="cri-o://8aa53bcf517e462fd9a64d48de52fd9484d3a2861af550e380a93ce6b9d97ec1" gracePeriod=30 Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.133216 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="sg-core" containerID="cri-o://84e16224ae48b3f3b23576debccd8b560aa3741a55ce07cc2d3079aac5f187a7" gracePeriod=30 Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.133266 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="ceilometer-notification-agent" containerID="cri-o://d0e8cb77705a9d74725495026ec255cb2b9da5d4bbcb6afcec32fd945fc118b9" gracePeriod=30 Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.150226 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.195:3000/\": EOF" Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.696552 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fqt9l"] Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.698510 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.710732 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fqt9l"] Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.854227 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ef0ad29-171d-4c53-81b2-835cb89bb447-utilities\") pod \"redhat-operators-fqt9l\" (UID: \"7ef0ad29-171d-4c53-81b2-835cb89bb447\") " pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.854622 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ef0ad29-171d-4c53-81b2-835cb89bb447-catalog-content\") pod \"redhat-operators-fqt9l\" (UID: \"7ef0ad29-171d-4c53-81b2-835cb89bb447\") " pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.854796 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czttz\" (UniqueName: \"kubernetes.io/projected/7ef0ad29-171d-4c53-81b2-835cb89bb447-kube-api-access-czttz\") pod \"redhat-operators-fqt9l\" (UID: \"7ef0ad29-171d-4c53-81b2-835cb89bb447\") " pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.956398 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czttz\" (UniqueName: \"kubernetes.io/projected/7ef0ad29-171d-4c53-81b2-835cb89bb447-kube-api-access-czttz\") pod \"redhat-operators-fqt9l\" (UID: \"7ef0ad29-171d-4c53-81b2-835cb89bb447\") " pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.956515 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ef0ad29-171d-4c53-81b2-835cb89bb447-utilities\") pod \"redhat-operators-fqt9l\" (UID: \"7ef0ad29-171d-4c53-81b2-835cb89bb447\") " pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.956594 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ef0ad29-171d-4c53-81b2-835cb89bb447-catalog-content\") pod \"redhat-operators-fqt9l\" (UID: \"7ef0ad29-171d-4c53-81b2-835cb89bb447\") " pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.957184 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ef0ad29-171d-4c53-81b2-835cb89bb447-catalog-content\") pod \"redhat-operators-fqt9l\" (UID: \"7ef0ad29-171d-4c53-81b2-835cb89bb447\") " pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:17:51 crc kubenswrapper[4806]: I1204 04:17:51.957785 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ef0ad29-171d-4c53-81b2-835cb89bb447-utilities\") pod \"redhat-operators-fqt9l\" (UID: \"7ef0ad29-171d-4c53-81b2-835cb89bb447\") " pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:17:52 crc kubenswrapper[4806]: I1204 04:17:52.000466 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czttz\" (UniqueName: \"kubernetes.io/projected/7ef0ad29-171d-4c53-81b2-835cb89bb447-kube-api-access-czttz\") pod \"redhat-operators-fqt9l\" (UID: \"7ef0ad29-171d-4c53-81b2-835cb89bb447\") " pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:17:52 crc kubenswrapper[4806]: I1204 04:17:52.022586 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:17:52 crc kubenswrapper[4806]: I1204 04:17:52.060538 4806 generic.go:334] "Generic (PLEG): container finished" podID="528ecdc7-e743-41de-b060-5aec0be99851" containerID="8aa53bcf517e462fd9a64d48de52fd9484d3a2861af550e380a93ce6b9d97ec1" exitCode=0 Dec 04 04:17:52 crc kubenswrapper[4806]: I1204 04:17:52.060573 4806 generic.go:334] "Generic (PLEG): container finished" podID="528ecdc7-e743-41de-b060-5aec0be99851" containerID="84e16224ae48b3f3b23576debccd8b560aa3741a55ce07cc2d3079aac5f187a7" exitCode=2 Dec 04 04:17:52 crc kubenswrapper[4806]: I1204 04:17:52.060582 4806 generic.go:334] "Generic (PLEG): container finished" podID="528ecdc7-e743-41de-b060-5aec0be99851" containerID="34d312119bbdf650410a1fadafe3de601e99be71ecd1ee5371fb4a820e14d05f" exitCode=0 Dec 04 04:17:52 crc kubenswrapper[4806]: I1204 04:17:52.060603 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"528ecdc7-e743-41de-b060-5aec0be99851","Type":"ContainerDied","Data":"8aa53bcf517e462fd9a64d48de52fd9484d3a2861af550e380a93ce6b9d97ec1"} Dec 04 04:17:52 crc kubenswrapper[4806]: I1204 04:17:52.060626 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"528ecdc7-e743-41de-b060-5aec0be99851","Type":"ContainerDied","Data":"84e16224ae48b3f3b23576debccd8b560aa3741a55ce07cc2d3079aac5f187a7"} Dec 04 04:17:52 crc kubenswrapper[4806]: I1204 04:17:52.060635 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"528ecdc7-e743-41de-b060-5aec0be99851","Type":"ContainerDied","Data":"34d312119bbdf650410a1fadafe3de601e99be71ecd1ee5371fb4a820e14d05f"} Dec 04 04:17:52 crc kubenswrapper[4806]: I1204 04:17:52.571312 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fqt9l"] Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.069827 4806 generic.go:334] "Generic (PLEG): container finished" podID="7ef0ad29-171d-4c53-81b2-835cb89bb447" containerID="78dd474001f95604fac94394a6eca9b052c3d6607374c3f6dac365531d9ddce8" exitCode=0 Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.069918 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fqt9l" event={"ID":"7ef0ad29-171d-4c53-81b2-835cb89bb447","Type":"ContainerDied","Data":"78dd474001f95604fac94394a6eca9b052c3d6607374c3f6dac365531d9ddce8"} Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.070156 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fqt9l" event={"ID":"7ef0ad29-171d-4c53-81b2-835cb89bb447","Type":"ContainerStarted","Data":"8314e79a60ddebfc2b41cad599777b5a1ee2c2af17e4fcb416f6aaada7cd47ee"} Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.461451 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.513049 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.714960 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.811770 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-logs\") pod \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\" (UID: \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\") " Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.811905 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ztqp\" (UniqueName: \"kubernetes.io/projected/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-kube-api-access-4ztqp\") pod \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\" (UID: \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\") " Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.812084 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-combined-ca-bundle\") pod \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\" (UID: \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\") " Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.812117 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-config-data\") pod \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\" (UID: \"1354ec28-6b19-4f91-9b52-fbcfd83b6fce\") " Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.812455 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-logs" (OuterVolumeSpecName: "logs") pod "1354ec28-6b19-4f91-9b52-fbcfd83b6fce" (UID: "1354ec28-6b19-4f91-9b52-fbcfd83b6fce"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.824465 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-kube-api-access-4ztqp" (OuterVolumeSpecName: "kube-api-access-4ztqp") pod "1354ec28-6b19-4f91-9b52-fbcfd83b6fce" (UID: "1354ec28-6b19-4f91-9b52-fbcfd83b6fce"). InnerVolumeSpecName "kube-api-access-4ztqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.884852 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-config-data" (OuterVolumeSpecName: "config-data") pod "1354ec28-6b19-4f91-9b52-fbcfd83b6fce" (UID: "1354ec28-6b19-4f91-9b52-fbcfd83b6fce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.900372 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1354ec28-6b19-4f91-9b52-fbcfd83b6fce" (UID: "1354ec28-6b19-4f91-9b52-fbcfd83b6fce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.914034 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ztqp\" (UniqueName: \"kubernetes.io/projected/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-kube-api-access-4ztqp\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.914080 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.914094 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:53 crc kubenswrapper[4806]: I1204 04:17:53.914106 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1354ec28-6b19-4f91-9b52-fbcfd83b6fce-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.084404 4806 generic.go:334] "Generic (PLEG): container finished" podID="1354ec28-6b19-4f91-9b52-fbcfd83b6fce" containerID="df2bcdcd0af1a361583a7328f1abb4da9b4bda4bc14ec666e73c34e16f5e6dc2" exitCode=0 Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.084849 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1354ec28-6b19-4f91-9b52-fbcfd83b6fce","Type":"ContainerDied","Data":"df2bcdcd0af1a361583a7328f1abb4da9b4bda4bc14ec666e73c34e16f5e6dc2"} Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.084887 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1354ec28-6b19-4f91-9b52-fbcfd83b6fce","Type":"ContainerDied","Data":"53efd9e51bb5305bdf228205a677ed461f457379fe877f752b84c3b820d8ff3a"} Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.084908 4806 scope.go:117] "RemoveContainer" containerID="df2bcdcd0af1a361583a7328f1abb4da9b4bda4bc14ec666e73c34e16f5e6dc2" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.085100 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.092148 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fqt9l" event={"ID":"7ef0ad29-171d-4c53-81b2-835cb89bb447","Type":"ContainerStarted","Data":"51ea19e4ad77466aeaf40b3c61f197f2aac1f290c6dd90692d340b0ef09cb7b2"} Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.120461 4806 scope.go:117] "RemoveContainer" containerID="73b3aa0ec02f6571b04650e56fb636fc2ccacc3fc01c56d6299de8613ca357ca" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.143881 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.153241 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.174315 4806 scope.go:117] "RemoveContainer" containerID="df2bcdcd0af1a361583a7328f1abb4da9b4bda4bc14ec666e73c34e16f5e6dc2" Dec 04 04:17:54 crc kubenswrapper[4806]: E1204 04:17:54.175845 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df2bcdcd0af1a361583a7328f1abb4da9b4bda4bc14ec666e73c34e16f5e6dc2\": container with ID starting with df2bcdcd0af1a361583a7328f1abb4da9b4bda4bc14ec666e73c34e16f5e6dc2 not found: ID does not exist" containerID="df2bcdcd0af1a361583a7328f1abb4da9b4bda4bc14ec666e73c34e16f5e6dc2" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.175893 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df2bcdcd0af1a361583a7328f1abb4da9b4bda4bc14ec666e73c34e16f5e6dc2"} err="failed to get container status \"df2bcdcd0af1a361583a7328f1abb4da9b4bda4bc14ec666e73c34e16f5e6dc2\": rpc error: code = NotFound desc = could not find container \"df2bcdcd0af1a361583a7328f1abb4da9b4bda4bc14ec666e73c34e16f5e6dc2\": container with ID starting with df2bcdcd0af1a361583a7328f1abb4da9b4bda4bc14ec666e73c34e16f5e6dc2 not found: ID does not exist" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.175940 4806 scope.go:117] "RemoveContainer" containerID="73b3aa0ec02f6571b04650e56fb636fc2ccacc3fc01c56d6299de8613ca357ca" Dec 04 04:17:54 crc kubenswrapper[4806]: E1204 04:17:54.176318 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73b3aa0ec02f6571b04650e56fb636fc2ccacc3fc01c56d6299de8613ca357ca\": container with ID starting with 73b3aa0ec02f6571b04650e56fb636fc2ccacc3fc01c56d6299de8613ca357ca not found: ID does not exist" containerID="73b3aa0ec02f6571b04650e56fb636fc2ccacc3fc01c56d6299de8613ca357ca" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.176339 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73b3aa0ec02f6571b04650e56fb636fc2ccacc3fc01c56d6299de8613ca357ca"} err="failed to get container status \"73b3aa0ec02f6571b04650e56fb636fc2ccacc3fc01c56d6299de8613ca357ca\": rpc error: code = NotFound desc = could not find container \"73b3aa0ec02f6571b04650e56fb636fc2ccacc3fc01c56d6299de8613ca357ca\": container with ID starting with 73b3aa0ec02f6571b04650e56fb636fc2ccacc3fc01c56d6299de8613ca357ca not found: ID does not exist" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.189525 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 04:17:54 crc kubenswrapper[4806]: E1204 04:17:54.189997 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1354ec28-6b19-4f91-9b52-fbcfd83b6fce" containerName="nova-api-log" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.190018 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="1354ec28-6b19-4f91-9b52-fbcfd83b6fce" containerName="nova-api-log" Dec 04 04:17:54 crc kubenswrapper[4806]: E1204 04:17:54.190038 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1354ec28-6b19-4f91-9b52-fbcfd83b6fce" containerName="nova-api-api" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.190045 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="1354ec28-6b19-4f91-9b52-fbcfd83b6fce" containerName="nova-api-api" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.190220 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="1354ec28-6b19-4f91-9b52-fbcfd83b6fce" containerName="nova-api-log" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.190244 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="1354ec28-6b19-4f91-9b52-fbcfd83b6fce" containerName="nova-api-api" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.191279 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.195847 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.196086 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.200025 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.247580 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.338549 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8kmv\" (UniqueName: \"kubernetes.io/projected/0d4a86f2-ea6b-4918-b3d9-1b2501139825-kube-api-access-t8kmv\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.339038 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-public-tls-certs\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.339088 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.339365 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.339480 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-config-data\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.339537 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d4a86f2-ea6b-4918-b3d9-1b2501139825-logs\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.341759 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.377629 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.442617 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8kmv\" (UniqueName: \"kubernetes.io/projected/0d4a86f2-ea6b-4918-b3d9-1b2501139825-kube-api-access-t8kmv\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.442677 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-public-tls-certs\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.442705 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.442765 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.442805 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-config-data\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.442827 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d4a86f2-ea6b-4918-b3d9-1b2501139825-logs\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.443294 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d4a86f2-ea6b-4918-b3d9-1b2501139825-logs\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.448326 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-config-data\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.450603 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-public-tls-certs\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.451002 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.457091 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.462813 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8kmv\" (UniqueName: \"kubernetes.io/projected/0d4a86f2-ea6b-4918-b3d9-1b2501139825-kube-api-access-t8kmv\") pod \"nova-api-0\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " pod="openstack/nova-api-0" Dec 04 04:17:54 crc kubenswrapper[4806]: I1204 04:17:54.524984 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.122762 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 04 04:17:55 crc kubenswrapper[4806]: W1204 04:17:55.223964 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d4a86f2_ea6b_4918_b3d9_1b2501139825.slice/crio-5d690a319c7b8d1a9ab0b212623e6f87cb6cc7f295cdce962b8891f3aaf3f86f WatchSource:0}: Error finding container 5d690a319c7b8d1a9ab0b212623e6f87cb6cc7f295cdce962b8891f3aaf3f86f: Status 404 returned error can't find the container with id 5d690a319c7b8d1a9ab0b212623e6f87cb6cc7f295cdce962b8891f3aaf3f86f Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.236206 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.365053 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-fgv9w"] Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.367172 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-fgv9w" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.376746 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.376859 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.386544 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-fgv9w"] Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.443108 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1354ec28-6b19-4f91-9b52-fbcfd83b6fce" path="/var/lib/kubelet/pods/1354ec28-6b19-4f91-9b52-fbcfd83b6fce/volumes" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.470118 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-scripts\") pod \"nova-cell1-cell-mapping-fgv9w\" (UID: \"d1f08876-fc00-4755-82ca-87352049c5d3\") " pod="openstack/nova-cell1-cell-mapping-fgv9w" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.470531 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-fgv9w\" (UID: \"d1f08876-fc00-4755-82ca-87352049c5d3\") " pod="openstack/nova-cell1-cell-mapping-fgv9w" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.470711 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-config-data\") pod \"nova-cell1-cell-mapping-fgv9w\" (UID: \"d1f08876-fc00-4755-82ca-87352049c5d3\") " pod="openstack/nova-cell1-cell-mapping-fgv9w" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.470944 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b8m8\" (UniqueName: \"kubernetes.io/projected/d1f08876-fc00-4755-82ca-87352049c5d3-kube-api-access-2b8m8\") pod \"nova-cell1-cell-mapping-fgv9w\" (UID: \"d1f08876-fc00-4755-82ca-87352049c5d3\") " pod="openstack/nova-cell1-cell-mapping-fgv9w" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.573213 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-fgv9w\" (UID: \"d1f08876-fc00-4755-82ca-87352049c5d3\") " pod="openstack/nova-cell1-cell-mapping-fgv9w" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.573903 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-config-data\") pod \"nova-cell1-cell-mapping-fgv9w\" (UID: \"d1f08876-fc00-4755-82ca-87352049c5d3\") " pod="openstack/nova-cell1-cell-mapping-fgv9w" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.574066 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b8m8\" (UniqueName: \"kubernetes.io/projected/d1f08876-fc00-4755-82ca-87352049c5d3-kube-api-access-2b8m8\") pod \"nova-cell1-cell-mapping-fgv9w\" (UID: \"d1f08876-fc00-4755-82ca-87352049c5d3\") " pod="openstack/nova-cell1-cell-mapping-fgv9w" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.574532 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-scripts\") pod \"nova-cell1-cell-mapping-fgv9w\" (UID: \"d1f08876-fc00-4755-82ca-87352049c5d3\") " pod="openstack/nova-cell1-cell-mapping-fgv9w" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.583321 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-config-data\") pod \"nova-cell1-cell-mapping-fgv9w\" (UID: \"d1f08876-fc00-4755-82ca-87352049c5d3\") " pod="openstack/nova-cell1-cell-mapping-fgv9w" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.609330 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b8m8\" (UniqueName: \"kubernetes.io/projected/d1f08876-fc00-4755-82ca-87352049c5d3-kube-api-access-2b8m8\") pod \"nova-cell1-cell-mapping-fgv9w\" (UID: \"d1f08876-fc00-4755-82ca-87352049c5d3\") " pod="openstack/nova-cell1-cell-mapping-fgv9w" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.622116 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-fgv9w\" (UID: \"d1f08876-fc00-4755-82ca-87352049c5d3\") " pod="openstack/nova-cell1-cell-mapping-fgv9w" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.623732 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-scripts\") pod \"nova-cell1-cell-mapping-fgv9w\" (UID: \"d1f08876-fc00-4755-82ca-87352049c5d3\") " pod="openstack/nova-cell1-cell-mapping-fgv9w" Dec 04 04:17:55 crc kubenswrapper[4806]: I1204 04:17:55.715711 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-fgv9w" Dec 04 04:17:56 crc kubenswrapper[4806]: I1204 04:17:56.126551 4806 generic.go:334] "Generic (PLEG): container finished" podID="528ecdc7-e743-41de-b060-5aec0be99851" containerID="d0e8cb77705a9d74725495026ec255cb2b9da5d4bbcb6afcec32fd945fc118b9" exitCode=0 Dec 04 04:17:56 crc kubenswrapper[4806]: I1204 04:17:56.127131 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"528ecdc7-e743-41de-b060-5aec0be99851","Type":"ContainerDied","Data":"d0e8cb77705a9d74725495026ec255cb2b9da5d4bbcb6afcec32fd945fc118b9"} Dec 04 04:17:56 crc kubenswrapper[4806]: I1204 04:17:56.131039 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d4a86f2-ea6b-4918-b3d9-1b2501139825","Type":"ContainerStarted","Data":"c9454e542d562159bccff3f1899236b38a6799396b34b6e6f3920108a81a3590"} Dec 04 04:17:56 crc kubenswrapper[4806]: I1204 04:17:56.131098 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d4a86f2-ea6b-4918-b3d9-1b2501139825","Type":"ContainerStarted","Data":"5d690a319c7b8d1a9ab0b212623e6f87cb6cc7f295cdce962b8891f3aaf3f86f"} Dec 04 04:17:56 crc kubenswrapper[4806]: W1204 04:17:56.477444 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1f08876_fc00_4755_82ca_87352049c5d3.slice/crio-444b76702d36f20a80f4c13a07a40c0d27d384e4b2edadf935a895f55d2ac80e WatchSource:0}: Error finding container 444b76702d36f20a80f4c13a07a40c0d27d384e4b2edadf935a895f55d2ac80e: Status 404 returned error can't find the container with id 444b76702d36f20a80f4c13a07a40c0d27d384e4b2edadf935a895f55d2ac80e Dec 04 04:17:56 crc kubenswrapper[4806]: I1204 04:17:56.487230 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-fgv9w"] Dec 04 04:17:56 crc kubenswrapper[4806]: I1204 04:17:56.902373 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.030886 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqfsw\" (UniqueName: \"kubernetes.io/projected/528ecdc7-e743-41de-b060-5aec0be99851-kube-api-access-qqfsw\") pod \"528ecdc7-e743-41de-b060-5aec0be99851\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.031084 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-ceilometer-tls-certs\") pod \"528ecdc7-e743-41de-b060-5aec0be99851\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.031135 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-config-data\") pod \"528ecdc7-e743-41de-b060-5aec0be99851\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.031346 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-combined-ca-bundle\") pod \"528ecdc7-e743-41de-b060-5aec0be99851\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.031444 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-sg-core-conf-yaml\") pod \"528ecdc7-e743-41de-b060-5aec0be99851\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.031524 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/528ecdc7-e743-41de-b060-5aec0be99851-run-httpd\") pod \"528ecdc7-e743-41de-b060-5aec0be99851\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.031566 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-scripts\") pod \"528ecdc7-e743-41de-b060-5aec0be99851\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.031597 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/528ecdc7-e743-41de-b060-5aec0be99851-log-httpd\") pod \"528ecdc7-e743-41de-b060-5aec0be99851\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.033109 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/528ecdc7-e743-41de-b060-5aec0be99851-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "528ecdc7-e743-41de-b060-5aec0be99851" (UID: "528ecdc7-e743-41de-b060-5aec0be99851"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.033542 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/528ecdc7-e743-41de-b060-5aec0be99851-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "528ecdc7-e743-41de-b060-5aec0be99851" (UID: "528ecdc7-e743-41de-b060-5aec0be99851"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.067320 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-scripts" (OuterVolumeSpecName: "scripts") pod "528ecdc7-e743-41de-b060-5aec0be99851" (UID: "528ecdc7-e743-41de-b060-5aec0be99851"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.067487 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/528ecdc7-e743-41de-b060-5aec0be99851-kube-api-access-qqfsw" (OuterVolumeSpecName: "kube-api-access-qqfsw") pod "528ecdc7-e743-41de-b060-5aec0be99851" (UID: "528ecdc7-e743-41de-b060-5aec0be99851"). InnerVolumeSpecName "kube-api-access-qqfsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.090092 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "528ecdc7-e743-41de-b060-5aec0be99851" (UID: "528ecdc7-e743-41de-b060-5aec0be99851"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.132786 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "528ecdc7-e743-41de-b060-5aec0be99851" (UID: "528ecdc7-e743-41de-b060-5aec0be99851"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.133915 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-ceilometer-tls-certs\") pod \"528ecdc7-e743-41de-b060-5aec0be99851\" (UID: \"528ecdc7-e743-41de-b060-5aec0be99851\") " Dec 04 04:17:57 crc kubenswrapper[4806]: W1204 04:17:57.134095 4806 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/528ecdc7-e743-41de-b060-5aec0be99851/volumes/kubernetes.io~secret/ceilometer-tls-certs Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.134113 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "528ecdc7-e743-41de-b060-5aec0be99851" (UID: "528ecdc7-e743-41de-b060-5aec0be99851"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.134616 4806 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.134637 4806 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/528ecdc7-e743-41de-b060-5aec0be99851-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.134646 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.134656 4806 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/528ecdc7-e743-41de-b060-5aec0be99851-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.134665 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqfsw\" (UniqueName: \"kubernetes.io/projected/528ecdc7-e743-41de-b060-5aec0be99851-kube-api-access-qqfsw\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.134675 4806 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.143090 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "528ecdc7-e743-41de-b060-5aec0be99851" (UID: "528ecdc7-e743-41de-b060-5aec0be99851"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.146256 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"528ecdc7-e743-41de-b060-5aec0be99851","Type":"ContainerDied","Data":"cff92425fe243adcc23e9b96d02fac4a75c4f3e2e5b03b9e3675d15758072684"} Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.146278 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.146313 4806 scope.go:117] "RemoveContainer" containerID="8aa53bcf517e462fd9a64d48de52fd9484d3a2861af550e380a93ce6b9d97ec1" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.153666 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d4a86f2-ea6b-4918-b3d9-1b2501139825","Type":"ContainerStarted","Data":"eb9935cdaf7951ef4430eef2664a79cef99dc46934821aabdc5488452efcab4a"} Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.166278 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-fgv9w" event={"ID":"d1f08876-fc00-4755-82ca-87352049c5d3","Type":"ContainerStarted","Data":"bf97c3e4038b06bfe8cd8862506302e6218b0048b8d8a9acd66a8c02e156fedc"} Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.166355 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-fgv9w" event={"ID":"d1f08876-fc00-4755-82ca-87352049c5d3","Type":"ContainerStarted","Data":"444b76702d36f20a80f4c13a07a40c0d27d384e4b2edadf935a895f55d2ac80e"} Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.196613 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.196594808 podStartE2EDuration="3.196594808s" podCreationTimestamp="2025-12-04 04:17:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:17:57.183454333 +0000 UTC m=+1392.041967261" watchObservedRunningTime="2025-12-04 04:17:57.196594808 +0000 UTC m=+1392.055107736" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.210151 4806 scope.go:117] "RemoveContainer" containerID="84e16224ae48b3f3b23576debccd8b560aa3741a55ce07cc2d3079aac5f187a7" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.232869 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-fgv9w" podStartSLOduration=2.232824635 podStartE2EDuration="2.232824635s" podCreationTimestamp="2025-12-04 04:17:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:17:57.201246055 +0000 UTC m=+1392.059758993" watchObservedRunningTime="2025-12-04 04:17:57.232824635 +0000 UTC m=+1392.091337583" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.237459 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.242888 4806 scope.go:117] "RemoveContainer" containerID="d0e8cb77705a9d74725495026ec255cb2b9da5d4bbcb6afcec32fd945fc118b9" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.278077 4806 scope.go:117] "RemoveContainer" containerID="34d312119bbdf650410a1fadafe3de601e99be71ecd1ee5371fb4a820e14d05f" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.307270 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-config-data" (OuterVolumeSpecName: "config-data") pod "528ecdc7-e743-41de-b060-5aec0be99851" (UID: "528ecdc7-e743-41de-b060-5aec0be99851"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.339042 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/528ecdc7-e743-41de-b060-5aec0be99851-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.491225 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.505118 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.520613 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:17:57 crc kubenswrapper[4806]: E1204 04:17:57.521325 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="sg-core" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.521417 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="sg-core" Dec 04 04:17:57 crc kubenswrapper[4806]: E1204 04:17:57.521491 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="ceilometer-notification-agent" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.521548 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="ceilometer-notification-agent" Dec 04 04:17:57 crc kubenswrapper[4806]: E1204 04:17:57.521615 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="ceilometer-central-agent" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.521681 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="ceilometer-central-agent" Dec 04 04:17:57 crc kubenswrapper[4806]: E1204 04:17:57.521745 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="proxy-httpd" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.521806 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="proxy-httpd" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.522065 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="proxy-httpd" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.522150 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="ceilometer-notification-agent" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.522222 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="sg-core" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.522298 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="528ecdc7-e743-41de-b060-5aec0be99851" containerName="ceilometer-central-agent" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.524103 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.530139 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.530604 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.530833 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.542419 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.601131 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.644197 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-run-httpd\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.644282 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s52pt\" (UniqueName: \"kubernetes.io/projected/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-kube-api-access-s52pt\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.644358 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-log-httpd\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.644378 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.644570 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.644693 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.644785 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-config-data\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.644986 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-scripts\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.679154 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-7zc27"] Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.679415 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-7zc27" podUID="733e0561-2895-4aad-a809-40333bf7120d" containerName="dnsmasq-dns" containerID="cri-o://7fe5631ecec872cd67648b893e68ec39600c044a04eccd576076793dac79abb3" gracePeriod=10 Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.751752 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-run-httpd\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.751848 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s52pt\" (UniqueName: \"kubernetes.io/projected/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-kube-api-access-s52pt\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.751912 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-log-httpd\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.751961 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.752009 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.752050 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.752092 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-config-data\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.752190 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-scripts\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.752573 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-log-httpd\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.753835 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-run-httpd\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.757887 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-scripts\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.762436 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.763323 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.768751 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-config-data\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.769428 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.788221 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s52pt\" (UniqueName: \"kubernetes.io/projected/56b1c86b-9508-4ca9-a2b8-9a496875c9e8-kube-api-access-s52pt\") pod \"ceilometer-0\" (UID: \"56b1c86b-9508-4ca9-a2b8-9a496875c9e8\") " pod="openstack/ceilometer-0" Dec 04 04:17:57 crc kubenswrapper[4806]: I1204 04:17:57.878526 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.216485 4806 generic.go:334] "Generic (PLEG): container finished" podID="7ef0ad29-171d-4c53-81b2-835cb89bb447" containerID="51ea19e4ad77466aeaf40b3c61f197f2aac1f290c6dd90692d340b0ef09cb7b2" exitCode=0 Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.216759 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fqt9l" event={"ID":"7ef0ad29-171d-4c53-81b2-835cb89bb447","Type":"ContainerDied","Data":"51ea19e4ad77466aeaf40b3c61f197f2aac1f290c6dd90692d340b0ef09cb7b2"} Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.223918 4806 generic.go:334] "Generic (PLEG): container finished" podID="733e0561-2895-4aad-a809-40333bf7120d" containerID="7fe5631ecec872cd67648b893e68ec39600c044a04eccd576076793dac79abb3" exitCode=0 Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.224237 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-7zc27" event={"ID":"733e0561-2895-4aad-a809-40333bf7120d","Type":"ContainerDied","Data":"7fe5631ecec872cd67648b893e68ec39600c044a04eccd576076793dac79abb3"} Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.287204 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.413502 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7d8bd6c96d-4sqhf" Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.562328 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bc4f74d8d-xj2xn"] Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.573715 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.716676 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-dns-svc\") pod \"733e0561-2895-4aad-a809-40333bf7120d\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.716737 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-config\") pod \"733e0561-2895-4aad-a809-40333bf7120d\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.716826 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-ovsdbserver-sb\") pod \"733e0561-2895-4aad-a809-40333bf7120d\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.716918 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-dns-swift-storage-0\") pod \"733e0561-2895-4aad-a809-40333bf7120d\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.716982 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-956ww\" (UniqueName: \"kubernetes.io/projected/733e0561-2895-4aad-a809-40333bf7120d-kube-api-access-956ww\") pod \"733e0561-2895-4aad-a809-40333bf7120d\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.717030 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-ovsdbserver-nb\") pod \"733e0561-2895-4aad-a809-40333bf7120d\" (UID: \"733e0561-2895-4aad-a809-40333bf7120d\") " Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.762055 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/733e0561-2895-4aad-a809-40333bf7120d-kube-api-access-956ww" (OuterVolumeSpecName: "kube-api-access-956ww") pod "733e0561-2895-4aad-a809-40333bf7120d" (UID: "733e0561-2895-4aad-a809-40333bf7120d"). InnerVolumeSpecName "kube-api-access-956ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.840382 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-956ww\" (UniqueName: \"kubernetes.io/projected/733e0561-2895-4aad-a809-40333bf7120d-kube-api-access-956ww\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.864711 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.937529 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "733e0561-2895-4aad-a809-40333bf7120d" (UID: "733e0561-2895-4aad-a809-40333bf7120d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:17:58 crc kubenswrapper[4806]: I1204 04:17:58.942063 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:58.997864 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "733e0561-2895-4aad-a809-40333bf7120d" (UID: "733e0561-2895-4aad-a809-40333bf7120d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.044393 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.070558 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "733e0561-2895-4aad-a809-40333bf7120d" (UID: "733e0561-2895-4aad-a809-40333bf7120d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.117084 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "733e0561-2895-4aad-a809-40333bf7120d" (UID: "733e0561-2895-4aad-a809-40333bf7120d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.126347 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-config" (OuterVolumeSpecName: "config") pod "733e0561-2895-4aad-a809-40333bf7120d" (UID: "733e0561-2895-4aad-a809-40333bf7120d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.146440 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.146478 4806 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.146491 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/733e0561-2895-4aad-a809-40333bf7120d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.285064 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56b1c86b-9508-4ca9-a2b8-9a496875c9e8","Type":"ContainerStarted","Data":"64411392590a463448a5435af329bb811fa4a7ba327effa4fe681d93a5cd850c"} Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.304010 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-7zc27" Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.304068 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-7zc27" event={"ID":"733e0561-2895-4aad-a809-40333bf7120d","Type":"ContainerDied","Data":"6801d4478ae4ac246fa054db026e7225daaa33e6c49ee5bf5d2635a24c6da3ab"} Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.304112 4806 scope.go:117] "RemoveContainer" containerID="7fe5631ecec872cd67648b893e68ec39600c044a04eccd576076793dac79abb3" Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.305323 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bc4f74d8d-xj2xn" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon-log" containerID="cri-o://1b52080a09d2ec89dad35352b7fbd08759432c499eaa43fddc058c1d3aa391e5" gracePeriod=30 Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.305510 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bc4f74d8d-xj2xn" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" containerID="cri-o://93f795e689362ccdc64196f9d3fe1246fe19535562972e9c4bbb63a9dccd7f18" gracePeriod=30 Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.361608 4806 scope.go:117] "RemoveContainer" containerID="92af52f3015c1fa0dbd735bdf3c206842c8abbeed44a7a6e3efdbded9c3654fe" Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.373994 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-7zc27"] Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.390847 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-7zc27"] Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.438893 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="528ecdc7-e743-41de-b060-5aec0be99851" path="/var/lib/kubelet/pods/528ecdc7-e743-41de-b060-5aec0be99851/volumes" Dec 04 04:17:59 crc kubenswrapper[4806]: I1204 04:17:59.440200 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="733e0561-2895-4aad-a809-40333bf7120d" path="/var/lib/kubelet/pods/733e0561-2895-4aad-a809-40333bf7120d/volumes" Dec 04 04:18:00 crc kubenswrapper[4806]: I1204 04:18:00.354186 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fqt9l" event={"ID":"7ef0ad29-171d-4c53-81b2-835cb89bb447","Type":"ContainerStarted","Data":"48345361ee8dd7c1e6e071ffe0ceef90106aecdeebb98e8002ac94ec06595023"} Dec 04 04:18:00 crc kubenswrapper[4806]: I1204 04:18:00.367151 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56b1c86b-9508-4ca9-a2b8-9a496875c9e8","Type":"ContainerStarted","Data":"c49547aad06b657a09fc934441ad54f84e74ce6a545ab3689af993fb8b6870c4"} Dec 04 04:18:00 crc kubenswrapper[4806]: I1204 04:18:00.382144 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fqt9l" podStartSLOduration=3.497271551 podStartE2EDuration="9.382125666s" podCreationTimestamp="2025-12-04 04:17:51 +0000 UTC" firstStartedPulling="2025-12-04 04:17:53.071663267 +0000 UTC m=+1387.930176185" lastFinishedPulling="2025-12-04 04:17:58.956517372 +0000 UTC m=+1393.815030300" observedRunningTime="2025-12-04 04:18:00.381163105 +0000 UTC m=+1395.239676043" watchObservedRunningTime="2025-12-04 04:18:00.382125666 +0000 UTC m=+1395.240638594" Dec 04 04:18:01 crc kubenswrapper[4806]: I1204 04:18:01.389410 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56b1c86b-9508-4ca9-a2b8-9a496875c9e8","Type":"ContainerStarted","Data":"d0eeb83a205a9c79a97159aeb54d46753c220570896f46154a3f8fc28e757b81"} Dec 04 04:18:02 crc kubenswrapper[4806]: I1204 04:18:02.023588 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:18:02 crc kubenswrapper[4806]: I1204 04:18:02.023717 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:18:02 crc kubenswrapper[4806]: I1204 04:18:02.402673 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56b1c86b-9508-4ca9-a2b8-9a496875c9e8","Type":"ContainerStarted","Data":"209fb0e554e9fd41c5cba4a784423cafe34b3396abc15d57ea5f9666928dc167"} Dec 04 04:18:02 crc kubenswrapper[4806]: I1204 04:18:02.514228 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7bc4f74d8d-xj2xn" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:58780->10.217.0.143:8443: read: connection reset by peer" Dec 04 04:18:03 crc kubenswrapper[4806]: I1204 04:18:03.078394 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fqt9l" podUID="7ef0ad29-171d-4c53-81b2-835cb89bb447" containerName="registry-server" probeResult="failure" output=< Dec 04 04:18:03 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Dec 04 04:18:03 crc kubenswrapper[4806]: > Dec 04 04:18:03 crc kubenswrapper[4806]: I1204 04:18:03.419783 4806 generic.go:334] "Generic (PLEG): container finished" podID="2265394b-be59-4797-b54d-369e5a5624e0" containerID="93f795e689362ccdc64196f9d3fe1246fe19535562972e9c4bbb63a9dccd7f18" exitCode=0 Dec 04 04:18:03 crc kubenswrapper[4806]: I1204 04:18:03.419825 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc4f74d8d-xj2xn" event={"ID":"2265394b-be59-4797-b54d-369e5a5624e0","Type":"ContainerDied","Data":"93f795e689362ccdc64196f9d3fe1246fe19535562972e9c4bbb63a9dccd7f18"} Dec 04 04:18:03 crc kubenswrapper[4806]: I1204 04:18:03.419859 4806 scope.go:117] "RemoveContainer" containerID="40e0a10f18db6fa40112391f3276df8272d8be2efe99e6d7b0a5ae22be825bb3" Dec 04 04:18:04 crc kubenswrapper[4806]: I1204 04:18:04.430471 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"56b1c86b-9508-4ca9-a2b8-9a496875c9e8","Type":"ContainerStarted","Data":"a81ddd0e268f00c2c3f3f38a4d491d128c5864d4c0ce81d70132a6bd6e370afd"} Dec 04 04:18:04 crc kubenswrapper[4806]: I1204 04:18:04.430763 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 04:18:04 crc kubenswrapper[4806]: I1204 04:18:04.450332 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.043909101 podStartE2EDuration="7.450314251s" podCreationTimestamp="2025-12-04 04:17:57 +0000 UTC" firstStartedPulling="2025-12-04 04:17:58.918150118 +0000 UTC m=+1393.776663036" lastFinishedPulling="2025-12-04 04:18:03.324555258 +0000 UTC m=+1398.183068186" observedRunningTime="2025-12-04 04:18:04.448542975 +0000 UTC m=+1399.307055903" watchObservedRunningTime="2025-12-04 04:18:04.450314251 +0000 UTC m=+1399.308827179" Dec 04 04:18:04 crc kubenswrapper[4806]: I1204 04:18:04.526094 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 04:18:04 crc kubenswrapper[4806]: I1204 04:18:04.526814 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 04:18:05 crc kubenswrapper[4806]: I1204 04:18:05.540126 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0d4a86f2-ea6b-4918-b3d9-1b2501139825" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:18:05 crc kubenswrapper[4806]: I1204 04:18:05.540132 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0d4a86f2-ea6b-4918-b3d9-1b2501139825" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:18:06 crc kubenswrapper[4806]: I1204 04:18:06.450614 4806 generic.go:334] "Generic (PLEG): container finished" podID="d1f08876-fc00-4755-82ca-87352049c5d3" containerID="bf97c3e4038b06bfe8cd8862506302e6218b0048b8d8a9acd66a8c02e156fedc" exitCode=0 Dec 04 04:18:06 crc kubenswrapper[4806]: I1204 04:18:06.450669 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-fgv9w" event={"ID":"d1f08876-fc00-4755-82ca-87352049c5d3","Type":"ContainerDied","Data":"bf97c3e4038b06bfe8cd8862506302e6218b0048b8d8a9acd66a8c02e156fedc"} Dec 04 04:18:07 crc kubenswrapper[4806]: I1204 04:18:07.899715 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-fgv9w" Dec 04 04:18:07 crc kubenswrapper[4806]: I1204 04:18:07.936015 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-config-data\") pod \"d1f08876-fc00-4755-82ca-87352049c5d3\" (UID: \"d1f08876-fc00-4755-82ca-87352049c5d3\") " Dec 04 04:18:07 crc kubenswrapper[4806]: I1204 04:18:07.936121 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-combined-ca-bundle\") pod \"d1f08876-fc00-4755-82ca-87352049c5d3\" (UID: \"d1f08876-fc00-4755-82ca-87352049c5d3\") " Dec 04 04:18:07 crc kubenswrapper[4806]: I1204 04:18:07.936270 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-scripts\") pod \"d1f08876-fc00-4755-82ca-87352049c5d3\" (UID: \"d1f08876-fc00-4755-82ca-87352049c5d3\") " Dec 04 04:18:07 crc kubenswrapper[4806]: I1204 04:18:07.936321 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2b8m8\" (UniqueName: \"kubernetes.io/projected/d1f08876-fc00-4755-82ca-87352049c5d3-kube-api-access-2b8m8\") pod \"d1f08876-fc00-4755-82ca-87352049c5d3\" (UID: \"d1f08876-fc00-4755-82ca-87352049c5d3\") " Dec 04 04:18:07 crc kubenswrapper[4806]: I1204 04:18:07.952567 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-scripts" (OuterVolumeSpecName: "scripts") pod "d1f08876-fc00-4755-82ca-87352049c5d3" (UID: "d1f08876-fc00-4755-82ca-87352049c5d3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:07 crc kubenswrapper[4806]: I1204 04:18:07.965975 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1f08876-fc00-4755-82ca-87352049c5d3-kube-api-access-2b8m8" (OuterVolumeSpecName: "kube-api-access-2b8m8") pod "d1f08876-fc00-4755-82ca-87352049c5d3" (UID: "d1f08876-fc00-4755-82ca-87352049c5d3"). InnerVolumeSpecName "kube-api-access-2b8m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:18:07 crc kubenswrapper[4806]: I1204 04:18:07.993914 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1f08876-fc00-4755-82ca-87352049c5d3" (UID: "d1f08876-fc00-4755-82ca-87352049c5d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:07 crc kubenswrapper[4806]: I1204 04:18:07.996942 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-config-data" (OuterVolumeSpecName: "config-data") pod "d1f08876-fc00-4755-82ca-87352049c5d3" (UID: "d1f08876-fc00-4755-82ca-87352049c5d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:08 crc kubenswrapper[4806]: I1204 04:18:08.038546 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2b8m8\" (UniqueName: \"kubernetes.io/projected/d1f08876-fc00-4755-82ca-87352049c5d3-kube-api-access-2b8m8\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:08 crc kubenswrapper[4806]: I1204 04:18:08.038596 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:08 crc kubenswrapper[4806]: I1204 04:18:08.038612 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:08 crc kubenswrapper[4806]: I1204 04:18:08.038624 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1f08876-fc00-4755-82ca-87352049c5d3-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:08 crc kubenswrapper[4806]: I1204 04:18:08.468042 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-fgv9w" event={"ID":"d1f08876-fc00-4755-82ca-87352049c5d3","Type":"ContainerDied","Data":"444b76702d36f20a80f4c13a07a40c0d27d384e4b2edadf935a895f55d2ac80e"} Dec 04 04:18:08 crc kubenswrapper[4806]: I1204 04:18:08.468083 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="444b76702d36f20a80f4c13a07a40c0d27d384e4b2edadf935a895f55d2ac80e" Dec 04 04:18:08 crc kubenswrapper[4806]: I1204 04:18:08.468095 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-fgv9w" Dec 04 04:18:08 crc kubenswrapper[4806]: I1204 04:18:08.668877 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 04:18:08 crc kubenswrapper[4806]: I1204 04:18:08.669121 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="77484080-5dac-49d3-b5c7-96aefe955966" containerName="nova-scheduler-scheduler" containerID="cri-o://8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6" gracePeriod=30 Dec 04 04:18:08 crc kubenswrapper[4806]: I1204 04:18:08.685063 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:18:08 crc kubenswrapper[4806]: I1204 04:18:08.685364 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="417de6e8-85e9-47d5-843a-c44096913ccb" containerName="nova-metadata-log" containerID="cri-o://d1874375f111276dce66e7b66c976108a8db4ebfd21470a75921ac1b2011db2e" gracePeriod=30 Dec 04 04:18:08 crc kubenswrapper[4806]: I1204 04:18:08.685447 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="417de6e8-85e9-47d5-843a-c44096913ccb" containerName="nova-metadata-metadata" containerID="cri-o://fdd0d0aa5b3c5b4c7be455ca09c59eada2a19044b1bae433a1366338176d2220" gracePeriod=30 Dec 04 04:18:08 crc kubenswrapper[4806]: I1204 04:18:08.699999 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:18:08 crc kubenswrapper[4806]: I1204 04:18:08.700524 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0d4a86f2-ea6b-4918-b3d9-1b2501139825" containerName="nova-api-log" containerID="cri-o://c9454e542d562159bccff3f1899236b38a6799396b34b6e6f3920108a81a3590" gracePeriod=30 Dec 04 04:18:08 crc kubenswrapper[4806]: I1204 04:18:08.700707 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0d4a86f2-ea6b-4918-b3d9-1b2501139825" containerName="nova-api-api" containerID="cri-o://eb9935cdaf7951ef4430eef2664a79cef99dc46934821aabdc5488452efcab4a" gracePeriod=30 Dec 04 04:18:09 crc kubenswrapper[4806]: I1204 04:18:09.494297 4806 generic.go:334] "Generic (PLEG): container finished" podID="0d4a86f2-ea6b-4918-b3d9-1b2501139825" containerID="c9454e542d562159bccff3f1899236b38a6799396b34b6e6f3920108a81a3590" exitCode=143 Dec 04 04:18:09 crc kubenswrapper[4806]: I1204 04:18:09.494380 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d4a86f2-ea6b-4918-b3d9-1b2501139825","Type":"ContainerDied","Data":"c9454e542d562159bccff3f1899236b38a6799396b34b6e6f3920108a81a3590"} Dec 04 04:18:09 crc kubenswrapper[4806]: I1204 04:18:09.504184 4806 generic.go:334] "Generic (PLEG): container finished" podID="417de6e8-85e9-47d5-843a-c44096913ccb" containerID="d1874375f111276dce66e7b66c976108a8db4ebfd21470a75921ac1b2011db2e" exitCode=143 Dec 04 04:18:09 crc kubenswrapper[4806]: I1204 04:18:09.504237 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"417de6e8-85e9-47d5-843a-c44096913ccb","Type":"ContainerDied","Data":"d1874375f111276dce66e7b66c976108a8db4ebfd21470a75921ac1b2011db2e"} Dec 04 04:18:09 crc kubenswrapper[4806]: I1204 04:18:09.943596 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7bc4f74d8d-xj2xn" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 04 04:18:11 crc kubenswrapper[4806]: I1204 04:18:11.863438 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="417de6e8-85e9-47d5-843a-c44096913ccb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": read tcp 10.217.0.2:48854->10.217.0.192:8775: read: connection reset by peer" Dec 04 04:18:11 crc kubenswrapper[4806]: I1204 04:18:11.863590 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="417de6e8-85e9-47d5-843a-c44096913ccb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.192:8775/\": read tcp 10.217.0.2:48856->10.217.0.192:8775: read: connection reset by peer" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.447788 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.461226 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.537872 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-nova-metadata-tls-certs\") pod \"417de6e8-85e9-47d5-843a-c44096913ccb\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.537941 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-config-data\") pod \"417de6e8-85e9-47d5-843a-c44096913ccb\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.537985 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-public-tls-certs\") pod \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.538687 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/417de6e8-85e9-47d5-843a-c44096913ccb-logs\") pod \"417de6e8-85e9-47d5-843a-c44096913ccb\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.538748 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-config-data\") pod \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.538777 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-internal-tls-certs\") pod \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.538801 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fwrs\" (UniqueName: \"kubernetes.io/projected/417de6e8-85e9-47d5-843a-c44096913ccb-kube-api-access-8fwrs\") pod \"417de6e8-85e9-47d5-843a-c44096913ccb\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.538818 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8kmv\" (UniqueName: \"kubernetes.io/projected/0d4a86f2-ea6b-4918-b3d9-1b2501139825-kube-api-access-t8kmv\") pod \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.538874 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-combined-ca-bundle\") pod \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.538894 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-combined-ca-bundle\") pod \"417de6e8-85e9-47d5-843a-c44096913ccb\" (UID: \"417de6e8-85e9-47d5-843a-c44096913ccb\") " Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.538981 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d4a86f2-ea6b-4918-b3d9-1b2501139825-logs\") pod \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\" (UID: \"0d4a86f2-ea6b-4918-b3d9-1b2501139825\") " Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.541730 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/417de6e8-85e9-47d5-843a-c44096913ccb-logs" (OuterVolumeSpecName: "logs") pod "417de6e8-85e9-47d5-843a-c44096913ccb" (UID: "417de6e8-85e9-47d5-843a-c44096913ccb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.544399 4806 generic.go:334] "Generic (PLEG): container finished" podID="0d4a86f2-ea6b-4918-b3d9-1b2501139825" containerID="eb9935cdaf7951ef4430eef2664a79cef99dc46934821aabdc5488452efcab4a" exitCode=0 Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.544833 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d4a86f2-ea6b-4918-b3d9-1b2501139825","Type":"ContainerDied","Data":"eb9935cdaf7951ef4430eef2664a79cef99dc46934821aabdc5488452efcab4a"} Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.544874 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d4a86f2-ea6b-4918-b3d9-1b2501139825","Type":"ContainerDied","Data":"5d690a319c7b8d1a9ab0b212623e6f87cb6cc7f295cdce962b8891f3aaf3f86f"} Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.544776 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.544891 4806 scope.go:117] "RemoveContainer" containerID="eb9935cdaf7951ef4430eef2664a79cef99dc46934821aabdc5488452efcab4a" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.555065 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d4a86f2-ea6b-4918-b3d9-1b2501139825-kube-api-access-t8kmv" (OuterVolumeSpecName: "kube-api-access-t8kmv") pod "0d4a86f2-ea6b-4918-b3d9-1b2501139825" (UID: "0d4a86f2-ea6b-4918-b3d9-1b2501139825"). InnerVolumeSpecName "kube-api-access-t8kmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.555751 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d4a86f2-ea6b-4918-b3d9-1b2501139825-logs" (OuterVolumeSpecName: "logs") pod "0d4a86f2-ea6b-4918-b3d9-1b2501139825" (UID: "0d4a86f2-ea6b-4918-b3d9-1b2501139825"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.558221 4806 generic.go:334] "Generic (PLEG): container finished" podID="417de6e8-85e9-47d5-843a-c44096913ccb" containerID="fdd0d0aa5b3c5b4c7be455ca09c59eada2a19044b1bae433a1366338176d2220" exitCode=0 Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.558268 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"417de6e8-85e9-47d5-843a-c44096913ccb","Type":"ContainerDied","Data":"fdd0d0aa5b3c5b4c7be455ca09c59eada2a19044b1bae433a1366338176d2220"} Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.558313 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"417de6e8-85e9-47d5-843a-c44096913ccb","Type":"ContainerDied","Data":"9d700145d0bb43156fc9df127784d31545e71cbfd9e588b9b03f34a79b7b305b"} Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.558372 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.566046 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/417de6e8-85e9-47d5-843a-c44096913ccb-kube-api-access-8fwrs" (OuterVolumeSpecName: "kube-api-access-8fwrs") pod "417de6e8-85e9-47d5-843a-c44096913ccb" (UID: "417de6e8-85e9-47d5-843a-c44096913ccb"). InnerVolumeSpecName "kube-api-access-8fwrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.593247 4806 scope.go:117] "RemoveContainer" containerID="c9454e542d562159bccff3f1899236b38a6799396b34b6e6f3920108a81a3590" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.610974 4806 scope.go:117] "RemoveContainer" containerID="eb9935cdaf7951ef4430eef2664a79cef99dc46934821aabdc5488452efcab4a" Dec 04 04:18:12 crc kubenswrapper[4806]: E1204 04:18:12.611615 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb9935cdaf7951ef4430eef2664a79cef99dc46934821aabdc5488452efcab4a\": container with ID starting with eb9935cdaf7951ef4430eef2664a79cef99dc46934821aabdc5488452efcab4a not found: ID does not exist" containerID="eb9935cdaf7951ef4430eef2664a79cef99dc46934821aabdc5488452efcab4a" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.611671 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb9935cdaf7951ef4430eef2664a79cef99dc46934821aabdc5488452efcab4a"} err="failed to get container status \"eb9935cdaf7951ef4430eef2664a79cef99dc46934821aabdc5488452efcab4a\": rpc error: code = NotFound desc = could not find container \"eb9935cdaf7951ef4430eef2664a79cef99dc46934821aabdc5488452efcab4a\": container with ID starting with eb9935cdaf7951ef4430eef2664a79cef99dc46934821aabdc5488452efcab4a not found: ID does not exist" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.611702 4806 scope.go:117] "RemoveContainer" containerID="c9454e542d562159bccff3f1899236b38a6799396b34b6e6f3920108a81a3590" Dec 04 04:18:12 crc kubenswrapper[4806]: E1204 04:18:12.612085 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9454e542d562159bccff3f1899236b38a6799396b34b6e6f3920108a81a3590\": container with ID starting with c9454e542d562159bccff3f1899236b38a6799396b34b6e6f3920108a81a3590 not found: ID does not exist" containerID="c9454e542d562159bccff3f1899236b38a6799396b34b6e6f3920108a81a3590" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.612112 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9454e542d562159bccff3f1899236b38a6799396b34b6e6f3920108a81a3590"} err="failed to get container status \"c9454e542d562159bccff3f1899236b38a6799396b34b6e6f3920108a81a3590\": rpc error: code = NotFound desc = could not find container \"c9454e542d562159bccff3f1899236b38a6799396b34b6e6f3920108a81a3590\": container with ID starting with c9454e542d562159bccff3f1899236b38a6799396b34b6e6f3920108a81a3590 not found: ID does not exist" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.612129 4806 scope.go:117] "RemoveContainer" containerID="fdd0d0aa5b3c5b4c7be455ca09c59eada2a19044b1bae433a1366338176d2220" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.623568 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d4a86f2-ea6b-4918-b3d9-1b2501139825" (UID: "0d4a86f2-ea6b-4918-b3d9-1b2501139825"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.635865 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-config-data" (OuterVolumeSpecName: "config-data") pod "0d4a86f2-ea6b-4918-b3d9-1b2501139825" (UID: "0d4a86f2-ea6b-4918-b3d9-1b2501139825"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.638914 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "417de6e8-85e9-47d5-843a-c44096913ccb" (UID: "417de6e8-85e9-47d5-843a-c44096913ccb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.642014 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/417de6e8-85e9-47d5-843a-c44096913ccb-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.642266 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.642367 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fwrs\" (UniqueName: \"kubernetes.io/projected/417de6e8-85e9-47d5-843a-c44096913ccb-kube-api-access-8fwrs\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.642457 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8kmv\" (UniqueName: \"kubernetes.io/projected/0d4a86f2-ea6b-4918-b3d9-1b2501139825-kube-api-access-t8kmv\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.642547 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.642699 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.642842 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d4a86f2-ea6b-4918-b3d9-1b2501139825-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.655053 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-config-data" (OuterVolumeSpecName: "config-data") pod "417de6e8-85e9-47d5-843a-c44096913ccb" (UID: "417de6e8-85e9-47d5-843a-c44096913ccb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.658367 4806 scope.go:117] "RemoveContainer" containerID="d1874375f111276dce66e7b66c976108a8db4ebfd21470a75921ac1b2011db2e" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.682702 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0d4a86f2-ea6b-4918-b3d9-1b2501139825" (UID: "0d4a86f2-ea6b-4918-b3d9-1b2501139825"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.695144 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0d4a86f2-ea6b-4918-b3d9-1b2501139825" (UID: "0d4a86f2-ea6b-4918-b3d9-1b2501139825"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.701101 4806 scope.go:117] "RemoveContainer" containerID="fdd0d0aa5b3c5b4c7be455ca09c59eada2a19044b1bae433a1366338176d2220" Dec 04 04:18:12 crc kubenswrapper[4806]: E1204 04:18:12.702407 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdd0d0aa5b3c5b4c7be455ca09c59eada2a19044b1bae433a1366338176d2220\": container with ID starting with fdd0d0aa5b3c5b4c7be455ca09c59eada2a19044b1bae433a1366338176d2220 not found: ID does not exist" containerID="fdd0d0aa5b3c5b4c7be455ca09c59eada2a19044b1bae433a1366338176d2220" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.702435 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdd0d0aa5b3c5b4c7be455ca09c59eada2a19044b1bae433a1366338176d2220"} err="failed to get container status \"fdd0d0aa5b3c5b4c7be455ca09c59eada2a19044b1bae433a1366338176d2220\": rpc error: code = NotFound desc = could not find container \"fdd0d0aa5b3c5b4c7be455ca09c59eada2a19044b1bae433a1366338176d2220\": container with ID starting with fdd0d0aa5b3c5b4c7be455ca09c59eada2a19044b1bae433a1366338176d2220 not found: ID does not exist" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.702454 4806 scope.go:117] "RemoveContainer" containerID="d1874375f111276dce66e7b66c976108a8db4ebfd21470a75921ac1b2011db2e" Dec 04 04:18:12 crc kubenswrapper[4806]: E1204 04:18:12.703366 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1874375f111276dce66e7b66c976108a8db4ebfd21470a75921ac1b2011db2e\": container with ID starting with d1874375f111276dce66e7b66c976108a8db4ebfd21470a75921ac1b2011db2e not found: ID does not exist" containerID="d1874375f111276dce66e7b66c976108a8db4ebfd21470a75921ac1b2011db2e" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.703392 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1874375f111276dce66e7b66c976108a8db4ebfd21470a75921ac1b2011db2e"} err="failed to get container status \"d1874375f111276dce66e7b66c976108a8db4ebfd21470a75921ac1b2011db2e\": rpc error: code = NotFound desc = could not find container \"d1874375f111276dce66e7b66c976108a8db4ebfd21470a75921ac1b2011db2e\": container with ID starting with d1874375f111276dce66e7b66c976108a8db4ebfd21470a75921ac1b2011db2e not found: ID does not exist" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.723557 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "417de6e8-85e9-47d5-843a-c44096913ccb" (UID: "417de6e8-85e9-47d5-843a-c44096913ccb"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.746774 4806 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.746849 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/417de6e8-85e9-47d5-843a-c44096913ccb-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.746863 4806 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.746875 4806 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d4a86f2-ea6b-4918-b3d9-1b2501139825-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.907027 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.922993 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.945598 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:18:12 crc kubenswrapper[4806]: I1204 04:18:12.980556 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.004408 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 04:18:13 crc kubenswrapper[4806]: E1204 04:18:13.004898 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d4a86f2-ea6b-4918-b3d9-1b2501139825" containerName="nova-api-api" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.004937 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d4a86f2-ea6b-4918-b3d9-1b2501139825" containerName="nova-api-api" Dec 04 04:18:13 crc kubenswrapper[4806]: E1204 04:18:13.004959 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d4a86f2-ea6b-4918-b3d9-1b2501139825" containerName="nova-api-log" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.004967 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d4a86f2-ea6b-4918-b3d9-1b2501139825" containerName="nova-api-log" Dec 04 04:18:13 crc kubenswrapper[4806]: E1204 04:18:13.004976 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="417de6e8-85e9-47d5-843a-c44096913ccb" containerName="nova-metadata-log" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.004983 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="417de6e8-85e9-47d5-843a-c44096913ccb" containerName="nova-metadata-log" Dec 04 04:18:13 crc kubenswrapper[4806]: E1204 04:18:13.005008 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="417de6e8-85e9-47d5-843a-c44096913ccb" containerName="nova-metadata-metadata" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.005015 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="417de6e8-85e9-47d5-843a-c44096913ccb" containerName="nova-metadata-metadata" Dec 04 04:18:13 crc kubenswrapper[4806]: E1204 04:18:13.005033 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f08876-fc00-4755-82ca-87352049c5d3" containerName="nova-manage" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.005039 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f08876-fc00-4755-82ca-87352049c5d3" containerName="nova-manage" Dec 04 04:18:13 crc kubenswrapper[4806]: E1204 04:18:13.005049 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="733e0561-2895-4aad-a809-40333bf7120d" containerName="dnsmasq-dns" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.005055 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="733e0561-2895-4aad-a809-40333bf7120d" containerName="dnsmasq-dns" Dec 04 04:18:13 crc kubenswrapper[4806]: E1204 04:18:13.005063 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="733e0561-2895-4aad-a809-40333bf7120d" containerName="init" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.005069 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="733e0561-2895-4aad-a809-40333bf7120d" containerName="init" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.005330 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="417de6e8-85e9-47d5-843a-c44096913ccb" containerName="nova-metadata-metadata" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.005343 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d4a86f2-ea6b-4918-b3d9-1b2501139825" containerName="nova-api-log" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.005356 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="733e0561-2895-4aad-a809-40333bf7120d" containerName="dnsmasq-dns" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.005368 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1f08876-fc00-4755-82ca-87352049c5d3" containerName="nova-manage" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.005381 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="417de6e8-85e9-47d5-843a-c44096913ccb" containerName="nova-metadata-log" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.005396 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d4a86f2-ea6b-4918-b3d9-1b2501139825" containerName="nova-api-api" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.006582 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.009887 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.010077 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.010193 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.021567 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.035503 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.037436 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.039749 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.040153 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.077008 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.083986 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d149e6e-bdab-4137-a48a-27ed21680ab8-logs\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.084372 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d149e6e-bdab-4137-a48a-27ed21680ab8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.084499 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d149e6e-bdab-4137-a48a-27ed21680ab8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.084634 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d149e6e-bdab-4137-a48a-27ed21680ab8-config-data\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.085172 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq7db\" (UniqueName: \"kubernetes.io/projected/4d149e6e-bdab-4137-a48a-27ed21680ab8-kube-api-access-sq7db\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.085406 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d149e6e-bdab-4137-a48a-27ed21680ab8-public-tls-certs\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.084502 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fqt9l" podUID="7ef0ad29-171d-4c53-81b2-835cb89bb447" containerName="registry-server" probeResult="failure" output=< Dec 04 04:18:13 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Dec 04 04:18:13 crc kubenswrapper[4806]: > Dec 04 04:18:13 crc kubenswrapper[4806]: E1204 04:18:13.096856 4806 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6 is running failed: container process not found" containerID="8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 04:18:13 crc kubenswrapper[4806]: E1204 04:18:13.098464 4806 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6 is running failed: container process not found" containerID="8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 04:18:13 crc kubenswrapper[4806]: E1204 04:18:13.102618 4806 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6 is running failed: container process not found" containerID="8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 04:18:13 crc kubenswrapper[4806]: E1204 04:18:13.102869 4806 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="77484080-5dac-49d3-b5c7-96aefe955966" containerName="nova-scheduler-scheduler" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.189984 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d149e6e-bdab-4137-a48a-27ed21680ab8-public-tls-certs\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.190122 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d149e6e-bdab-4137-a48a-27ed21680ab8-logs\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.190158 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac-config-data\") pod \"nova-metadata-0\" (UID: \"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac\") " pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.190182 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac\") " pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.190209 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac-logs\") pod \"nova-metadata-0\" (UID: \"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac\") " pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.190246 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d149e6e-bdab-4137-a48a-27ed21680ab8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.190273 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac\") " pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.190306 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d149e6e-bdab-4137-a48a-27ed21680ab8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.190356 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d149e6e-bdab-4137-a48a-27ed21680ab8-config-data\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.190394 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj5dk\" (UniqueName: \"kubernetes.io/projected/c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac-kube-api-access-rj5dk\") pod \"nova-metadata-0\" (UID: \"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac\") " pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.190424 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq7db\" (UniqueName: \"kubernetes.io/projected/4d149e6e-bdab-4137-a48a-27ed21680ab8-kube-api-access-sq7db\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.195861 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d149e6e-bdab-4137-a48a-27ed21680ab8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.196433 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d149e6e-bdab-4137-a48a-27ed21680ab8-logs\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.198898 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d149e6e-bdab-4137-a48a-27ed21680ab8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.200566 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d149e6e-bdab-4137-a48a-27ed21680ab8-config-data\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.201220 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d149e6e-bdab-4137-a48a-27ed21680ab8-public-tls-certs\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.215420 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq7db\" (UniqueName: \"kubernetes.io/projected/4d149e6e-bdab-4137-a48a-27ed21680ab8-kube-api-access-sq7db\") pod \"nova-api-0\" (UID: \"4d149e6e-bdab-4137-a48a-27ed21680ab8\") " pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.292443 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac-logs\") pod \"nova-metadata-0\" (UID: \"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac\") " pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.292529 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac\") " pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.292593 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj5dk\" (UniqueName: \"kubernetes.io/projected/c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac-kube-api-access-rj5dk\") pod \"nova-metadata-0\" (UID: \"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac\") " pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.292721 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac-config-data\") pod \"nova-metadata-0\" (UID: \"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac\") " pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.292747 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac\") " pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.293015 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac-logs\") pod \"nova-metadata-0\" (UID: \"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac\") " pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.297129 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac\") " pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.298801 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac\") " pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.299639 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac-config-data\") pod \"nova-metadata-0\" (UID: \"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac\") " pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.316380 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj5dk\" (UniqueName: \"kubernetes.io/projected/c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac-kube-api-access-rj5dk\") pod \"nova-metadata-0\" (UID: \"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac\") " pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.379997 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.401714 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.415350 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.457041 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d4a86f2-ea6b-4918-b3d9-1b2501139825" path="/var/lib/kubelet/pods/0d4a86f2-ea6b-4918-b3d9-1b2501139825/volumes" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.463306 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="417de6e8-85e9-47d5-843a-c44096913ccb" path="/var/lib/kubelet/pods/417de6e8-85e9-47d5-843a-c44096913ccb/volumes" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.495609 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwqhx\" (UniqueName: \"kubernetes.io/projected/77484080-5dac-49d3-b5c7-96aefe955966-kube-api-access-gwqhx\") pod \"77484080-5dac-49d3-b5c7-96aefe955966\" (UID: \"77484080-5dac-49d3-b5c7-96aefe955966\") " Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.495737 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77484080-5dac-49d3-b5c7-96aefe955966-combined-ca-bundle\") pod \"77484080-5dac-49d3-b5c7-96aefe955966\" (UID: \"77484080-5dac-49d3-b5c7-96aefe955966\") " Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.495918 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77484080-5dac-49d3-b5c7-96aefe955966-config-data\") pod \"77484080-5dac-49d3-b5c7-96aefe955966\" (UID: \"77484080-5dac-49d3-b5c7-96aefe955966\") " Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.504324 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77484080-5dac-49d3-b5c7-96aefe955966-kube-api-access-gwqhx" (OuterVolumeSpecName: "kube-api-access-gwqhx") pod "77484080-5dac-49d3-b5c7-96aefe955966" (UID: "77484080-5dac-49d3-b5c7-96aefe955966"). InnerVolumeSpecName "kube-api-access-gwqhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.541340 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77484080-5dac-49d3-b5c7-96aefe955966-config-data" (OuterVolumeSpecName: "config-data") pod "77484080-5dac-49d3-b5c7-96aefe955966" (UID: "77484080-5dac-49d3-b5c7-96aefe955966"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.556099 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77484080-5dac-49d3-b5c7-96aefe955966-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77484080-5dac-49d3-b5c7-96aefe955966" (UID: "77484080-5dac-49d3-b5c7-96aefe955966"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.597799 4806 generic.go:334] "Generic (PLEG): container finished" podID="77484080-5dac-49d3-b5c7-96aefe955966" containerID="8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6" exitCode=0 Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.597865 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"77484080-5dac-49d3-b5c7-96aefe955966","Type":"ContainerDied","Data":"8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6"} Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.597892 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"77484080-5dac-49d3-b5c7-96aefe955966","Type":"ContainerDied","Data":"2bf2215d7fa6e06f3699823476db2070b86d4b36a6a0573e474781381c171a16"} Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.597907 4806 scope.go:117] "RemoveContainer" containerID="8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.598027 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.607799 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77484080-5dac-49d3-b5c7-96aefe955966-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.607831 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwqhx\" (UniqueName: \"kubernetes.io/projected/77484080-5dac-49d3-b5c7-96aefe955966-kube-api-access-gwqhx\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.607843 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77484080-5dac-49d3-b5c7-96aefe955966-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.668524 4806 scope.go:117] "RemoveContainer" containerID="8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6" Dec 04 04:18:13 crc kubenswrapper[4806]: E1204 04:18:13.671458 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6\": container with ID starting with 8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6 not found: ID does not exist" containerID="8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.671509 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6"} err="failed to get container status \"8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6\": rpc error: code = NotFound desc = could not find container \"8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6\": container with ID starting with 8e0448f6127c4f4485a024b50fc1b9c0195943031fe7f855aeaed3c63036ebb6 not found: ID does not exist" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.696036 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.708122 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.723313 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 04:18:13 crc kubenswrapper[4806]: E1204 04:18:13.723756 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77484080-5dac-49d3-b5c7-96aefe955966" containerName="nova-scheduler-scheduler" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.723775 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="77484080-5dac-49d3-b5c7-96aefe955966" containerName="nova-scheduler-scheduler" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.723993 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="77484080-5dac-49d3-b5c7-96aefe955966" containerName="nova-scheduler-scheduler" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.724735 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.726774 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.731778 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.810793 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32a7b4e-8881-437b-843e-54a6d1e5c11e-config-data\") pod \"nova-scheduler-0\" (UID: \"a32a7b4e-8881-437b-843e-54a6d1e5c11e\") " pod="openstack/nova-scheduler-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.810981 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw6fs\" (UniqueName: \"kubernetes.io/projected/a32a7b4e-8881-437b-843e-54a6d1e5c11e-kube-api-access-tw6fs\") pod \"nova-scheduler-0\" (UID: \"a32a7b4e-8881-437b-843e-54a6d1e5c11e\") " pod="openstack/nova-scheduler-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.811031 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32a7b4e-8881-437b-843e-54a6d1e5c11e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a32a7b4e-8881-437b-843e-54a6d1e5c11e\") " pod="openstack/nova-scheduler-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.913227 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw6fs\" (UniqueName: \"kubernetes.io/projected/a32a7b4e-8881-437b-843e-54a6d1e5c11e-kube-api-access-tw6fs\") pod \"nova-scheduler-0\" (UID: \"a32a7b4e-8881-437b-843e-54a6d1e5c11e\") " pod="openstack/nova-scheduler-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.913305 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32a7b4e-8881-437b-843e-54a6d1e5c11e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a32a7b4e-8881-437b-843e-54a6d1e5c11e\") " pod="openstack/nova-scheduler-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.913417 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32a7b4e-8881-437b-843e-54a6d1e5c11e-config-data\") pod \"nova-scheduler-0\" (UID: \"a32a7b4e-8881-437b-843e-54a6d1e5c11e\") " pod="openstack/nova-scheduler-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.923557 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a32a7b4e-8881-437b-843e-54a6d1e5c11e-config-data\") pod \"nova-scheduler-0\" (UID: \"a32a7b4e-8881-437b-843e-54a6d1e5c11e\") " pod="openstack/nova-scheduler-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.928502 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a32a7b4e-8881-437b-843e-54a6d1e5c11e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a32a7b4e-8881-437b-843e-54a6d1e5c11e\") " pod="openstack/nova-scheduler-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.934609 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw6fs\" (UniqueName: \"kubernetes.io/projected/a32a7b4e-8881-437b-843e-54a6d1e5c11e-kube-api-access-tw6fs\") pod \"nova-scheduler-0\" (UID: \"a32a7b4e-8881-437b-843e-54a6d1e5c11e\") " pod="openstack/nova-scheduler-0" Dec 04 04:18:13 crc kubenswrapper[4806]: I1204 04:18:13.990563 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 04:18:14 crc kubenswrapper[4806]: W1204 04:18:14.003776 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d149e6e_bdab_4137_a48a_27ed21680ab8.slice/crio-d560e980397779acb02007710e379258c9edb08500026c5190c6c0bdbbad43ad WatchSource:0}: Error finding container d560e980397779acb02007710e379258c9edb08500026c5190c6c0bdbbad43ad: Status 404 returned error can't find the container with id d560e980397779acb02007710e379258c9edb08500026c5190c6c0bdbbad43ad Dec 04 04:18:14 crc kubenswrapper[4806]: I1204 04:18:14.043613 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 04:18:14 crc kubenswrapper[4806]: I1204 04:18:14.074664 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 04:18:14 crc kubenswrapper[4806]: I1204 04:18:14.582048 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 04:18:14 crc kubenswrapper[4806]: W1204 04:18:14.586446 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda32a7b4e_8881_437b_843e_54a6d1e5c11e.slice/crio-dbdb88274d7d40c1d6c22ed90a9a862261855606466c739b3fc7ad5e5fd82954 WatchSource:0}: Error finding container dbdb88274d7d40c1d6c22ed90a9a862261855606466c739b3fc7ad5e5fd82954: Status 404 returned error can't find the container with id dbdb88274d7d40c1d6c22ed90a9a862261855606466c739b3fc7ad5e5fd82954 Dec 04 04:18:14 crc kubenswrapper[4806]: I1204 04:18:14.638505 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4d149e6e-bdab-4137-a48a-27ed21680ab8","Type":"ContainerStarted","Data":"5d798513f77b34456427225aa70d40dae9a652a9e15f0c1d22c89eee5300cec0"} Dec 04 04:18:14 crc kubenswrapper[4806]: I1204 04:18:14.638545 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4d149e6e-bdab-4137-a48a-27ed21680ab8","Type":"ContainerStarted","Data":"d560e980397779acb02007710e379258c9edb08500026c5190c6c0bdbbad43ad"} Dec 04 04:18:14 crc kubenswrapper[4806]: I1204 04:18:14.641744 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac","Type":"ContainerStarted","Data":"289d817a015fb33ed5c3ddaee1225ad75a2f24f6b0d658df6db7a064ca1aa14b"} Dec 04 04:18:14 crc kubenswrapper[4806]: I1204 04:18:14.641776 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac","Type":"ContainerStarted","Data":"355bd43d43508e6d6a5c8bf4afb2f072eebab134c8e4b9acb443a869538825e5"} Dec 04 04:18:14 crc kubenswrapper[4806]: I1204 04:18:14.641786 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac","Type":"ContainerStarted","Data":"45de75092c338b30b32a8945ff529aa2b406cd2a44573f91358fb0cfb8ce1fb8"} Dec 04 04:18:14 crc kubenswrapper[4806]: I1204 04:18:14.649298 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a32a7b4e-8881-437b-843e-54a6d1e5c11e","Type":"ContainerStarted","Data":"dbdb88274d7d40c1d6c22ed90a9a862261855606466c739b3fc7ad5e5fd82954"} Dec 04 04:18:14 crc kubenswrapper[4806]: I1204 04:18:14.675004 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.674979589 podStartE2EDuration="2.674979589s" podCreationTimestamp="2025-12-04 04:18:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:18:14.664678012 +0000 UTC m=+1409.523190960" watchObservedRunningTime="2025-12-04 04:18:14.674979589 +0000 UTC m=+1409.533492527" Dec 04 04:18:15 crc kubenswrapper[4806]: I1204 04:18:15.471773 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77484080-5dac-49d3-b5c7-96aefe955966" path="/var/lib/kubelet/pods/77484080-5dac-49d3-b5c7-96aefe955966/volumes" Dec 04 04:18:15 crc kubenswrapper[4806]: I1204 04:18:15.659413 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a32a7b4e-8881-437b-843e-54a6d1e5c11e","Type":"ContainerStarted","Data":"86ce4d36358dffca1e4864751065d610ae8596a72fd50cf16a4b6de2bbb7d7e8"} Dec 04 04:18:15 crc kubenswrapper[4806]: I1204 04:18:15.661354 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4d149e6e-bdab-4137-a48a-27ed21680ab8","Type":"ContainerStarted","Data":"7228a23e0dab7b1ef09ffab15ff51a3686aa1fc09529520bfd5bf42e0395a1b8"} Dec 04 04:18:15 crc kubenswrapper[4806]: I1204 04:18:15.682701 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.682683645 podStartE2EDuration="2.682683645s" podCreationTimestamp="2025-12-04 04:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:18:15.676910102 +0000 UTC m=+1410.535423040" watchObservedRunningTime="2025-12-04 04:18:15.682683645 +0000 UTC m=+1410.541196573" Dec 04 04:18:15 crc kubenswrapper[4806]: I1204 04:18:15.710525 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.710503825 podStartE2EDuration="3.710503825s" podCreationTimestamp="2025-12-04 04:18:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:18:15.698299728 +0000 UTC m=+1410.556812646" watchObservedRunningTime="2025-12-04 04:18:15.710503825 +0000 UTC m=+1410.569016753" Dec 04 04:18:18 crc kubenswrapper[4806]: I1204 04:18:18.403132 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 04:18:18 crc kubenswrapper[4806]: I1204 04:18:18.403199 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 04:18:19 crc kubenswrapper[4806]: I1204 04:18:19.044701 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 04:18:19 crc kubenswrapper[4806]: I1204 04:18:19.943795 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7bc4f74d8d-xj2xn" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 04 04:18:19 crc kubenswrapper[4806]: I1204 04:18:19.944494 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:18:22 crc kubenswrapper[4806]: I1204 04:18:22.082288 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:18:22 crc kubenswrapper[4806]: I1204 04:18:22.148377 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:18:22 crc kubenswrapper[4806]: I1204 04:18:22.901042 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fqt9l"] Dec 04 04:18:23 crc kubenswrapper[4806]: I1204 04:18:23.380870 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 04:18:23 crc kubenswrapper[4806]: I1204 04:18:23.380954 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 04:18:23 crc kubenswrapper[4806]: I1204 04:18:23.403311 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 04:18:23 crc kubenswrapper[4806]: I1204 04:18:23.403377 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 04:18:23 crc kubenswrapper[4806]: I1204 04:18:23.744964 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fqt9l" podUID="7ef0ad29-171d-4c53-81b2-835cb89bb447" containerName="registry-server" containerID="cri-o://48345361ee8dd7c1e6e071ffe0ceef90106aecdeebb98e8002ac94ec06595023" gracePeriod=2 Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.044819 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.089539 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.329418 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.402274 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4d149e6e-bdab-4137-a48a-27ed21680ab8" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.402312 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4d149e6e-bdab-4137-a48a-27ed21680ab8" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.415205 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.415205 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.435353 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ef0ad29-171d-4c53-81b2-835cb89bb447-catalog-content\") pod \"7ef0ad29-171d-4c53-81b2-835cb89bb447\" (UID: \"7ef0ad29-171d-4c53-81b2-835cb89bb447\") " Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.435485 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ef0ad29-171d-4c53-81b2-835cb89bb447-utilities\") pod \"7ef0ad29-171d-4c53-81b2-835cb89bb447\" (UID: \"7ef0ad29-171d-4c53-81b2-835cb89bb447\") " Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.435515 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czttz\" (UniqueName: \"kubernetes.io/projected/7ef0ad29-171d-4c53-81b2-835cb89bb447-kube-api-access-czttz\") pod \"7ef0ad29-171d-4c53-81b2-835cb89bb447\" (UID: \"7ef0ad29-171d-4c53-81b2-835cb89bb447\") " Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.436825 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ef0ad29-171d-4c53-81b2-835cb89bb447-utilities" (OuterVolumeSpecName: "utilities") pod "7ef0ad29-171d-4c53-81b2-835cb89bb447" (UID: "7ef0ad29-171d-4c53-81b2-835cb89bb447"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.459390 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ef0ad29-171d-4c53-81b2-835cb89bb447-kube-api-access-czttz" (OuterVolumeSpecName: "kube-api-access-czttz") pod "7ef0ad29-171d-4c53-81b2-835cb89bb447" (UID: "7ef0ad29-171d-4c53-81b2-835cb89bb447"). InnerVolumeSpecName "kube-api-access-czttz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.539324 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ef0ad29-171d-4c53-81b2-835cb89bb447-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.539374 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czttz\" (UniqueName: \"kubernetes.io/projected/7ef0ad29-171d-4c53-81b2-835cb89bb447-kube-api-access-czttz\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.574433 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ef0ad29-171d-4c53-81b2-835cb89bb447-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ef0ad29-171d-4c53-81b2-835cb89bb447" (UID: "7ef0ad29-171d-4c53-81b2-835cb89bb447"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.640896 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ef0ad29-171d-4c53-81b2-835cb89bb447-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.766131 4806 generic.go:334] "Generic (PLEG): container finished" podID="7ef0ad29-171d-4c53-81b2-835cb89bb447" containerID="48345361ee8dd7c1e6e071ffe0ceef90106aecdeebb98e8002ac94ec06595023" exitCode=0 Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.767034 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fqt9l" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.771143 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fqt9l" event={"ID":"7ef0ad29-171d-4c53-81b2-835cb89bb447","Type":"ContainerDied","Data":"48345361ee8dd7c1e6e071ffe0ceef90106aecdeebb98e8002ac94ec06595023"} Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.771203 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fqt9l" event={"ID":"7ef0ad29-171d-4c53-81b2-835cb89bb447","Type":"ContainerDied","Data":"8314e79a60ddebfc2b41cad599777b5a1ee2c2af17e4fcb416f6aaada7cd47ee"} Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.771225 4806 scope.go:117] "RemoveContainer" containerID="48345361ee8dd7c1e6e071ffe0ceef90106aecdeebb98e8002ac94ec06595023" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.810081 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fqt9l"] Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.811109 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.817470 4806 scope.go:117] "RemoveContainer" containerID="51ea19e4ad77466aeaf40b3c61f197f2aac1f290c6dd90692d340b0ef09cb7b2" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.818280 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fqt9l"] Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.862790 4806 scope.go:117] "RemoveContainer" containerID="78dd474001f95604fac94394a6eca9b052c3d6607374c3f6dac365531d9ddce8" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.892971 4806 scope.go:117] "RemoveContainer" containerID="48345361ee8dd7c1e6e071ffe0ceef90106aecdeebb98e8002ac94ec06595023" Dec 04 04:18:24 crc kubenswrapper[4806]: E1204 04:18:24.893387 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48345361ee8dd7c1e6e071ffe0ceef90106aecdeebb98e8002ac94ec06595023\": container with ID starting with 48345361ee8dd7c1e6e071ffe0ceef90106aecdeebb98e8002ac94ec06595023 not found: ID does not exist" containerID="48345361ee8dd7c1e6e071ffe0ceef90106aecdeebb98e8002ac94ec06595023" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.893417 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48345361ee8dd7c1e6e071ffe0ceef90106aecdeebb98e8002ac94ec06595023"} err="failed to get container status \"48345361ee8dd7c1e6e071ffe0ceef90106aecdeebb98e8002ac94ec06595023\": rpc error: code = NotFound desc = could not find container \"48345361ee8dd7c1e6e071ffe0ceef90106aecdeebb98e8002ac94ec06595023\": container with ID starting with 48345361ee8dd7c1e6e071ffe0ceef90106aecdeebb98e8002ac94ec06595023 not found: ID does not exist" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.893435 4806 scope.go:117] "RemoveContainer" containerID="51ea19e4ad77466aeaf40b3c61f197f2aac1f290c6dd90692d340b0ef09cb7b2" Dec 04 04:18:24 crc kubenswrapper[4806]: E1204 04:18:24.893631 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51ea19e4ad77466aeaf40b3c61f197f2aac1f290c6dd90692d340b0ef09cb7b2\": container with ID starting with 51ea19e4ad77466aeaf40b3c61f197f2aac1f290c6dd90692d340b0ef09cb7b2 not found: ID does not exist" containerID="51ea19e4ad77466aeaf40b3c61f197f2aac1f290c6dd90692d340b0ef09cb7b2" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.893655 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51ea19e4ad77466aeaf40b3c61f197f2aac1f290c6dd90692d340b0ef09cb7b2"} err="failed to get container status \"51ea19e4ad77466aeaf40b3c61f197f2aac1f290c6dd90692d340b0ef09cb7b2\": rpc error: code = NotFound desc = could not find container \"51ea19e4ad77466aeaf40b3c61f197f2aac1f290c6dd90692d340b0ef09cb7b2\": container with ID starting with 51ea19e4ad77466aeaf40b3c61f197f2aac1f290c6dd90692d340b0ef09cb7b2 not found: ID does not exist" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.893668 4806 scope.go:117] "RemoveContainer" containerID="78dd474001f95604fac94394a6eca9b052c3d6607374c3f6dac365531d9ddce8" Dec 04 04:18:24 crc kubenswrapper[4806]: E1204 04:18:24.893870 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78dd474001f95604fac94394a6eca9b052c3d6607374c3f6dac365531d9ddce8\": container with ID starting with 78dd474001f95604fac94394a6eca9b052c3d6607374c3f6dac365531d9ddce8 not found: ID does not exist" containerID="78dd474001f95604fac94394a6eca9b052c3d6607374c3f6dac365531d9ddce8" Dec 04 04:18:24 crc kubenswrapper[4806]: I1204 04:18:24.893902 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78dd474001f95604fac94394a6eca9b052c3d6607374c3f6dac365531d9ddce8"} err="failed to get container status \"78dd474001f95604fac94394a6eca9b052c3d6607374c3f6dac365531d9ddce8\": rpc error: code = NotFound desc = could not find container \"78dd474001f95604fac94394a6eca9b052c3d6607374c3f6dac365531d9ddce8\": container with ID starting with 78dd474001f95604fac94394a6eca9b052c3d6607374c3f6dac365531d9ddce8 not found: ID does not exist" Dec 04 04:18:25 crc kubenswrapper[4806]: I1204 04:18:25.436569 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ef0ad29-171d-4c53-81b2-835cb89bb447" path="/var/lib/kubelet/pods/7ef0ad29-171d-4c53-81b2-835cb89bb447/volumes" Dec 04 04:18:27 crc kubenswrapper[4806]: I1204 04:18:27.046986 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:18:27 crc kubenswrapper[4806]: I1204 04:18:27.047291 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:18:27 crc kubenswrapper[4806]: I1204 04:18:27.894548 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.763679 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.836352 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-combined-ca-bundle\") pod \"2265394b-be59-4797-b54d-369e5a5624e0\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.836464 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzlm9\" (UniqueName: \"kubernetes.io/projected/2265394b-be59-4797-b54d-369e5a5624e0-kube-api-access-kzlm9\") pod \"2265394b-be59-4797-b54d-369e5a5624e0\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.836515 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-horizon-secret-key\") pod \"2265394b-be59-4797-b54d-369e5a5624e0\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.836540 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2265394b-be59-4797-b54d-369e5a5624e0-logs\") pod \"2265394b-be59-4797-b54d-369e5a5624e0\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.836635 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-horizon-tls-certs\") pod \"2265394b-be59-4797-b54d-369e5a5624e0\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.836704 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2265394b-be59-4797-b54d-369e5a5624e0-config-data\") pod \"2265394b-be59-4797-b54d-369e5a5624e0\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.836775 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2265394b-be59-4797-b54d-369e5a5624e0-scripts\") pod \"2265394b-be59-4797-b54d-369e5a5624e0\" (UID: \"2265394b-be59-4797-b54d-369e5a5624e0\") " Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.839079 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2265394b-be59-4797-b54d-369e5a5624e0-logs" (OuterVolumeSpecName: "logs") pod "2265394b-be59-4797-b54d-369e5a5624e0" (UID: "2265394b-be59-4797-b54d-369e5a5624e0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.845044 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "2265394b-be59-4797-b54d-369e5a5624e0" (UID: "2265394b-be59-4797-b54d-369e5a5624e0"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.846374 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2265394b-be59-4797-b54d-369e5a5624e0-kube-api-access-kzlm9" (OuterVolumeSpecName: "kube-api-access-kzlm9") pod "2265394b-be59-4797-b54d-369e5a5624e0" (UID: "2265394b-be59-4797-b54d-369e5a5624e0"). InnerVolumeSpecName "kube-api-access-kzlm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.852728 4806 generic.go:334] "Generic (PLEG): container finished" podID="2265394b-be59-4797-b54d-369e5a5624e0" containerID="1b52080a09d2ec89dad35352b7fbd08759432c499eaa43fddc058c1d3aa391e5" exitCode=137 Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.852786 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bc4f74d8d-xj2xn" Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.852806 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc4f74d8d-xj2xn" event={"ID":"2265394b-be59-4797-b54d-369e5a5624e0","Type":"ContainerDied","Data":"1b52080a09d2ec89dad35352b7fbd08759432c499eaa43fddc058c1d3aa391e5"} Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.852841 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bc4f74d8d-xj2xn" event={"ID":"2265394b-be59-4797-b54d-369e5a5624e0","Type":"ContainerDied","Data":"b961042d0e3bd0d5b615befeef373210308ebe50776af5b705d242c124924f6f"} Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.852884 4806 scope.go:117] "RemoveContainer" containerID="93f795e689362ccdc64196f9d3fe1246fe19535562972e9c4bbb63a9dccd7f18" Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.909448 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2265394b-be59-4797-b54d-369e5a5624e0" (UID: "2265394b-be59-4797-b54d-369e5a5624e0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.910330 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2265394b-be59-4797-b54d-369e5a5624e0-config-data" (OuterVolumeSpecName: "config-data") pod "2265394b-be59-4797-b54d-369e5a5624e0" (UID: "2265394b-be59-4797-b54d-369e5a5624e0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.942187 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2265394b-be59-4797-b54d-369e5a5624e0-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.942230 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.942242 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzlm9\" (UniqueName: \"kubernetes.io/projected/2265394b-be59-4797-b54d-369e5a5624e0-kube-api-access-kzlm9\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.942252 4806 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.942261 4806 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2265394b-be59-4797-b54d-369e5a5624e0-logs\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.949881 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2265394b-be59-4797-b54d-369e5a5624e0-scripts" (OuterVolumeSpecName: "scripts") pod "2265394b-be59-4797-b54d-369e5a5624e0" (UID: "2265394b-be59-4797-b54d-369e5a5624e0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:18:29 crc kubenswrapper[4806]: I1204 04:18:29.990590 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "2265394b-be59-4797-b54d-369e5a5624e0" (UID: "2265394b-be59-4797-b54d-369e5a5624e0"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:30 crc kubenswrapper[4806]: I1204 04:18:30.044301 4806 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/2265394b-be59-4797-b54d-369e5a5624e0-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:30 crc kubenswrapper[4806]: I1204 04:18:30.044346 4806 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2265394b-be59-4797-b54d-369e5a5624e0-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:30 crc kubenswrapper[4806]: I1204 04:18:30.162280 4806 scope.go:117] "RemoveContainer" containerID="1b52080a09d2ec89dad35352b7fbd08759432c499eaa43fddc058c1d3aa391e5" Dec 04 04:18:30 crc kubenswrapper[4806]: I1204 04:18:30.191023 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bc4f74d8d-xj2xn"] Dec 04 04:18:30 crc kubenswrapper[4806]: I1204 04:18:30.193797 4806 scope.go:117] "RemoveContainer" containerID="93f795e689362ccdc64196f9d3fe1246fe19535562972e9c4bbb63a9dccd7f18" Dec 04 04:18:30 crc kubenswrapper[4806]: E1204 04:18:30.194557 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93f795e689362ccdc64196f9d3fe1246fe19535562972e9c4bbb63a9dccd7f18\": container with ID starting with 93f795e689362ccdc64196f9d3fe1246fe19535562972e9c4bbb63a9dccd7f18 not found: ID does not exist" containerID="93f795e689362ccdc64196f9d3fe1246fe19535562972e9c4bbb63a9dccd7f18" Dec 04 04:18:30 crc kubenswrapper[4806]: I1204 04:18:30.194609 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93f795e689362ccdc64196f9d3fe1246fe19535562972e9c4bbb63a9dccd7f18"} err="failed to get container status \"93f795e689362ccdc64196f9d3fe1246fe19535562972e9c4bbb63a9dccd7f18\": rpc error: code = NotFound desc = could not find container \"93f795e689362ccdc64196f9d3fe1246fe19535562972e9c4bbb63a9dccd7f18\": container with ID starting with 93f795e689362ccdc64196f9d3fe1246fe19535562972e9c4bbb63a9dccd7f18 not found: ID does not exist" Dec 04 04:18:30 crc kubenswrapper[4806]: I1204 04:18:30.194640 4806 scope.go:117] "RemoveContainer" containerID="1b52080a09d2ec89dad35352b7fbd08759432c499eaa43fddc058c1d3aa391e5" Dec 04 04:18:30 crc kubenswrapper[4806]: E1204 04:18:30.195405 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b52080a09d2ec89dad35352b7fbd08759432c499eaa43fddc058c1d3aa391e5\": container with ID starting with 1b52080a09d2ec89dad35352b7fbd08759432c499eaa43fddc058c1d3aa391e5 not found: ID does not exist" containerID="1b52080a09d2ec89dad35352b7fbd08759432c499eaa43fddc058c1d3aa391e5" Dec 04 04:18:30 crc kubenswrapper[4806]: I1204 04:18:30.195446 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b52080a09d2ec89dad35352b7fbd08759432c499eaa43fddc058c1d3aa391e5"} err="failed to get container status \"1b52080a09d2ec89dad35352b7fbd08759432c499eaa43fddc058c1d3aa391e5\": rpc error: code = NotFound desc = could not find container \"1b52080a09d2ec89dad35352b7fbd08759432c499eaa43fddc058c1d3aa391e5\": container with ID starting with 1b52080a09d2ec89dad35352b7fbd08759432c499eaa43fddc058c1d3aa391e5 not found: ID does not exist" Dec 04 04:18:30 crc kubenswrapper[4806]: I1204 04:18:30.199584 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7bc4f74d8d-xj2xn"] Dec 04 04:18:31 crc kubenswrapper[4806]: I1204 04:18:31.436523 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2265394b-be59-4797-b54d-369e5a5624e0" path="/var/lib/kubelet/pods/2265394b-be59-4797-b54d-369e5a5624e0/volumes" Dec 04 04:18:33 crc kubenswrapper[4806]: I1204 04:18:33.387652 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 04:18:33 crc kubenswrapper[4806]: I1204 04:18:33.388065 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 04:18:33 crc kubenswrapper[4806]: I1204 04:18:33.388563 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 04:18:33 crc kubenswrapper[4806]: I1204 04:18:33.388585 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 04:18:33 crc kubenswrapper[4806]: I1204 04:18:33.395956 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 04:18:33 crc kubenswrapper[4806]: I1204 04:18:33.397272 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 04:18:33 crc kubenswrapper[4806]: I1204 04:18:33.412271 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 04:18:33 crc kubenswrapper[4806]: I1204 04:18:33.412549 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 04:18:33 crc kubenswrapper[4806]: I1204 04:18:33.416995 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 04:18:33 crc kubenswrapper[4806]: I1204 04:18:33.420375 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 04:18:41 crc kubenswrapper[4806]: I1204 04:18:41.839363 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 04:18:42 crc kubenswrapper[4806]: I1204 04:18:42.690718 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 04:18:46 crc kubenswrapper[4806]: I1204 04:18:46.740076 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="0b7d7449-e532-4a32-bfcd-dc9f92767bd6" containerName="rabbitmq" containerID="cri-o://fa63f5e0190e9d9dd4e1c9e13a341601eb4661dec32b87ddf56918276f9bad62" gracePeriod=604796 Dec 04 04:18:47 crc kubenswrapper[4806]: I1204 04:18:47.372179 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="f18b8381-c1f9-4e48-9639-adfc5a406726" containerName="rabbitmq" containerID="cri-o://818f720223746c8f111996df6c232ef39f5651b1db629200e3810d1486707a6f" gracePeriod=604796 Dec 04 04:18:49 crc kubenswrapper[4806]: I1204 04:18:49.281681 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="f18b8381-c1f9-4e48-9639-adfc5a406726" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 04 04:18:49 crc kubenswrapper[4806]: I1204 04:18:49.690078 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="0b7d7449-e532-4a32-bfcd-dc9f92767bd6" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.228618 4806 generic.go:334] "Generic (PLEG): container finished" podID="0b7d7449-e532-4a32-bfcd-dc9f92767bd6" containerID="fa63f5e0190e9d9dd4e1c9e13a341601eb4661dec32b87ddf56918276f9bad62" exitCode=0 Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.229322 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0b7d7449-e532-4a32-bfcd-dc9f92767bd6","Type":"ContainerDied","Data":"fa63f5e0190e9d9dd4e1c9e13a341601eb4661dec32b87ddf56918276f9bad62"} Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.229349 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0b7d7449-e532-4a32-bfcd-dc9f92767bd6","Type":"ContainerDied","Data":"00831f055f546ac605c4cbaecc3d90db3164d4e16e66117fa45ee873406b097d"} Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.229359 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00831f055f546ac605c4cbaecc3d90db3164d4e16e66117fa45ee873406b097d" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.283057 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.319881 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-erlang-cookie-secret\") pod \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.319938 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7bc6\" (UniqueName: \"kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-kube-api-access-l7bc6\") pod \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.319962 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-plugins-conf\") pod \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.319999 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-plugins\") pod \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.320029 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-confd\") pod \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.320068 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-erlang-cookie\") pod \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.320091 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-pod-info\") pod \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.320110 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-server-conf\") pod \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.320141 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-config-data\") pod \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.320204 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-tls\") pod \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.322359 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0b7d7449-e532-4a32-bfcd-dc9f92767bd6" (UID: "0b7d7449-e532-4a32-bfcd-dc9f92767bd6"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.325203 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0b7d7449-e532-4a32-bfcd-dc9f92767bd6" (UID: "0b7d7449-e532-4a32-bfcd-dc9f92767bd6"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.325255 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0b7d7449-e532-4a32-bfcd-dc9f92767bd6" (UID: "0b7d7449-e532-4a32-bfcd-dc9f92767bd6"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.325748 4806 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.325767 4806 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.325780 4806 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.336451 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0b7d7449-e532-4a32-bfcd-dc9f92767bd6" (UID: "0b7d7449-e532-4a32-bfcd-dc9f92767bd6"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.336604 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0b7d7449-e532-4a32-bfcd-dc9f92767bd6" (UID: "0b7d7449-e532-4a32-bfcd-dc9f92767bd6"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.356240 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-pod-info" (OuterVolumeSpecName: "pod-info") pod "0b7d7449-e532-4a32-bfcd-dc9f92767bd6" (UID: "0b7d7449-e532-4a32-bfcd-dc9f92767bd6"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.358179 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-kube-api-access-l7bc6" (OuterVolumeSpecName: "kube-api-access-l7bc6") pod "0b7d7449-e532-4a32-bfcd-dc9f92767bd6" (UID: "0b7d7449-e532-4a32-bfcd-dc9f92767bd6"). InnerVolumeSpecName "kube-api-access-l7bc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.385547 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-config-data" (OuterVolumeSpecName: "config-data") pod "0b7d7449-e532-4a32-bfcd-dc9f92767bd6" (UID: "0b7d7449-e532-4a32-bfcd-dc9f92767bd6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.426491 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\" (UID: \"0b7d7449-e532-4a32-bfcd-dc9f92767bd6\") " Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.426857 4806 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.426868 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.426876 4806 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.426885 4806 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.426894 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7bc6\" (UniqueName: \"kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-kube-api-access-l7bc6\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.451260 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-server-conf" (OuterVolumeSpecName: "server-conf") pod "0b7d7449-e532-4a32-bfcd-dc9f92767bd6" (UID: "0b7d7449-e532-4a32-bfcd-dc9f92767bd6"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.459378 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "0b7d7449-e532-4a32-bfcd-dc9f92767bd6" (UID: "0b7d7449-e532-4a32-bfcd-dc9f92767bd6"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.522992 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0b7d7449-e532-4a32-bfcd-dc9f92767bd6" (UID: "0b7d7449-e532-4a32-bfcd-dc9f92767bd6"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.528069 4806 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.528106 4806 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.528119 4806 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0b7d7449-e532-4a32-bfcd-dc9f92767bd6-server-conf\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.551620 4806 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 04 04:18:53 crc kubenswrapper[4806]: I1204 04:18:53.631027 4806 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.299495 4806 generic.go:334] "Generic (PLEG): container finished" podID="f18b8381-c1f9-4e48-9639-adfc5a406726" containerID="818f720223746c8f111996df6c232ef39f5651b1db629200e3810d1486707a6f" exitCode=0 Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.299984 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.300308 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f18b8381-c1f9-4e48-9639-adfc5a406726","Type":"ContainerDied","Data":"818f720223746c8f111996df6c232ef39f5651b1db629200e3810d1486707a6f"} Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.300372 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f18b8381-c1f9-4e48-9639-adfc5a406726","Type":"ContainerDied","Data":"b820a3f39cc9a745c86cb6837feaa651ad78664699cb08440440e659544397fa"} Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.301688 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b820a3f39cc9a745c86cb6837feaa651ad78664699cb08440440e659544397fa" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.310784 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.397994 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.413092 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.435018 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 04:18:54 crc kubenswrapper[4806]: E1204 04:18:54.435526 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ef0ad29-171d-4c53-81b2-835cb89bb447" containerName="extract-utilities" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.435548 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ef0ad29-171d-4c53-81b2-835cb89bb447" containerName="extract-utilities" Dec 04 04:18:54 crc kubenswrapper[4806]: E1204 04:18:54.435569 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b7d7449-e532-4a32-bfcd-dc9f92767bd6" containerName="setup-container" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.435582 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b7d7449-e532-4a32-bfcd-dc9f92767bd6" containerName="setup-container" Dec 04 04:18:54 crc kubenswrapper[4806]: E1204 04:18:54.435605 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ef0ad29-171d-4c53-81b2-835cb89bb447" containerName="extract-content" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.435613 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ef0ad29-171d-4c53-81b2-835cb89bb447" containerName="extract-content" Dec 04 04:18:54 crc kubenswrapper[4806]: E1204 04:18:54.435632 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.435641 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" Dec 04 04:18:54 crc kubenswrapper[4806]: E1204 04:18:54.435653 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.435662 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" Dec 04 04:18:54 crc kubenswrapper[4806]: E1204 04:18:54.435677 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon-log" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.435686 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon-log" Dec 04 04:18:54 crc kubenswrapper[4806]: E1204 04:18:54.435702 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f18b8381-c1f9-4e48-9639-adfc5a406726" containerName="setup-container" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.435709 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f18b8381-c1f9-4e48-9639-adfc5a406726" containerName="setup-container" Dec 04 04:18:54 crc kubenswrapper[4806]: E1204 04:18:54.435727 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f18b8381-c1f9-4e48-9639-adfc5a406726" containerName="rabbitmq" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.435736 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f18b8381-c1f9-4e48-9639-adfc5a406726" containerName="rabbitmq" Dec 04 04:18:54 crc kubenswrapper[4806]: E1204 04:18:54.435754 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b7d7449-e532-4a32-bfcd-dc9f92767bd6" containerName="rabbitmq" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.435762 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b7d7449-e532-4a32-bfcd-dc9f92767bd6" containerName="rabbitmq" Dec 04 04:18:54 crc kubenswrapper[4806]: E1204 04:18:54.435776 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ef0ad29-171d-4c53-81b2-835cb89bb447" containerName="registry-server" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.435784 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ef0ad29-171d-4c53-81b2-835cb89bb447" containerName="registry-server" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.436021 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f18b8381-c1f9-4e48-9639-adfc5a406726" containerName="rabbitmq" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.436055 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.436073 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.436088 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon-log" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.436101 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.436111 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b7d7449-e532-4a32-bfcd-dc9f92767bd6" containerName="rabbitmq" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.436123 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ef0ad29-171d-4c53-81b2-835cb89bb447" containerName="registry-server" Dec 04 04:18:54 crc kubenswrapper[4806]: E1204 04:18:54.436342 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.436351 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2265394b-be59-4797-b54d-369e5a5624e0" containerName="horizon" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.440486 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.443762 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.445535 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.445577 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.445742 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.446162 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.446315 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-xqlhc" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.446453 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.452452 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25bnr\" (UniqueName: \"kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-kube-api-access-25bnr\") pod \"f18b8381-c1f9-4e48-9639-adfc5a406726\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.452517 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-tls\") pod \"f18b8381-c1f9-4e48-9639-adfc5a406726\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.452627 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f18b8381-c1f9-4e48-9639-adfc5a406726-pod-info\") pod \"f18b8381-c1f9-4e48-9639-adfc5a406726\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.452664 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-erlang-cookie\") pod \"f18b8381-c1f9-4e48-9639-adfc5a406726\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.452816 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-plugins-conf\") pod \"f18b8381-c1f9-4e48-9639-adfc5a406726\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.452860 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"f18b8381-c1f9-4e48-9639-adfc5a406726\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.452891 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-plugins\") pod \"f18b8381-c1f9-4e48-9639-adfc5a406726\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.452972 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-config-data\") pod \"f18b8381-c1f9-4e48-9639-adfc5a406726\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.453014 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-server-conf\") pod \"f18b8381-c1f9-4e48-9639-adfc5a406726\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.453037 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f18b8381-c1f9-4e48-9639-adfc5a406726-erlang-cookie-secret\") pod \"f18b8381-c1f9-4e48-9639-adfc5a406726\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.453099 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-confd\") pod \"f18b8381-c1f9-4e48-9639-adfc5a406726\" (UID: \"f18b8381-c1f9-4e48-9639-adfc5a406726\") " Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.454258 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f18b8381-c1f9-4e48-9639-adfc5a406726" (UID: "f18b8381-c1f9-4e48-9639-adfc5a406726"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.457872 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f18b8381-c1f9-4e48-9639-adfc5a406726" (UID: "f18b8381-c1f9-4e48-9639-adfc5a406726"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.467463 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f18b8381-c1f9-4e48-9639-adfc5a406726" (UID: "f18b8381-c1f9-4e48-9639-adfc5a406726"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.468061 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f18b8381-c1f9-4e48-9639-adfc5a406726-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f18b8381-c1f9-4e48-9639-adfc5a406726" (UID: "f18b8381-c1f9-4e48-9639-adfc5a406726"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.470089 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.470731 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "f18b8381-c1f9-4e48-9639-adfc5a406726" (UID: "f18b8381-c1f9-4e48-9639-adfc5a406726"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.470912 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "f18b8381-c1f9-4e48-9639-adfc5a406726" (UID: "f18b8381-c1f9-4e48-9639-adfc5a406726"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.482129 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f18b8381-c1f9-4e48-9639-adfc5a406726-pod-info" (OuterVolumeSpecName: "pod-info") pod "f18b8381-c1f9-4e48-9639-adfc5a406726" (UID: "f18b8381-c1f9-4e48-9639-adfc5a406726"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.499158 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-kube-api-access-25bnr" (OuterVolumeSpecName: "kube-api-access-25bnr") pod "f18b8381-c1f9-4e48-9639-adfc5a406726" (UID: "f18b8381-c1f9-4e48-9639-adfc5a406726"). InnerVolumeSpecName "kube-api-access-25bnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.557444 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5a2a4a0f-f385-4fad-9825-2824f47dd8db-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.557510 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5a2a4a0f-f385-4fad-9825-2824f47dd8db-config-data\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.557544 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5a2a4a0f-f385-4fad-9825-2824f47dd8db-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.557614 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ff8f\" (UniqueName: \"kubernetes.io/projected/5a2a4a0f-f385-4fad-9825-2824f47dd8db-kube-api-access-4ff8f\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.557664 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5a2a4a0f-f385-4fad-9825-2824f47dd8db-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.557709 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.557954 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5a2a4a0f-f385-4fad-9825-2824f47dd8db-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.558042 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5a2a4a0f-f385-4fad-9825-2824f47dd8db-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.558133 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5a2a4a0f-f385-4fad-9825-2824f47dd8db-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.558203 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5a2a4a0f-f385-4fad-9825-2824f47dd8db-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.558220 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5a2a4a0f-f385-4fad-9825-2824f47dd8db-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.558447 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25bnr\" (UniqueName: \"kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-kube-api-access-25bnr\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.558478 4806 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.558491 4806 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f18b8381-c1f9-4e48-9639-adfc5a406726-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.558504 4806 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.558517 4806 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.558539 4806 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.558552 4806 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.558563 4806 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f18b8381-c1f9-4e48-9639-adfc5a406726-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.566775 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-config-data" (OuterVolumeSpecName: "config-data") pod "f18b8381-c1f9-4e48-9639-adfc5a406726" (UID: "f18b8381-c1f9-4e48-9639-adfc5a406726"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.613099 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-server-conf" (OuterVolumeSpecName: "server-conf") pod "f18b8381-c1f9-4e48-9639-adfc5a406726" (UID: "f18b8381-c1f9-4e48-9639-adfc5a406726"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.619465 4806 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.660642 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5a2a4a0f-f385-4fad-9825-2824f47dd8db-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.660686 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5a2a4a0f-f385-4fad-9825-2824f47dd8db-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.660730 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5a2a4a0f-f385-4fad-9825-2824f47dd8db-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.661004 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5a2a4a0f-f385-4fad-9825-2824f47dd8db-config-data\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.661027 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5a2a4a0f-f385-4fad-9825-2824f47dd8db-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.661058 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ff8f\" (UniqueName: \"kubernetes.io/projected/5a2a4a0f-f385-4fad-9825-2824f47dd8db-kube-api-access-4ff8f\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.661085 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5a2a4a0f-f385-4fad-9825-2824f47dd8db-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.661115 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.661156 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5a2a4a0f-f385-4fad-9825-2824f47dd8db-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.661185 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5a2a4a0f-f385-4fad-9825-2824f47dd8db-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.661213 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5a2a4a0f-f385-4fad-9825-2824f47dd8db-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.661260 4806 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.661270 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.661282 4806 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f18b8381-c1f9-4e48-9639-adfc5a406726-server-conf\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.661678 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5a2a4a0f-f385-4fad-9825-2824f47dd8db-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.663325 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5a2a4a0f-f385-4fad-9825-2824f47dd8db-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.663471 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.664094 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5a2a4a0f-f385-4fad-9825-2824f47dd8db-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.664331 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5a2a4a0f-f385-4fad-9825-2824f47dd8db-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.664841 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5a2a4a0f-f385-4fad-9825-2824f47dd8db-config-data\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.668170 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5a2a4a0f-f385-4fad-9825-2824f47dd8db-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.672628 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5a2a4a0f-f385-4fad-9825-2824f47dd8db-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.678191 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5a2a4a0f-f385-4fad-9825-2824f47dd8db-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.678568 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5a2a4a0f-f385-4fad-9825-2824f47dd8db-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.682525 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ff8f\" (UniqueName: \"kubernetes.io/projected/5a2a4a0f-f385-4fad-9825-2824f47dd8db-kube-api-access-4ff8f\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.716410 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f18b8381-c1f9-4e48-9639-adfc5a406726" (UID: "f18b8381-c1f9-4e48-9639-adfc5a406726"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.744463 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"5a2a4a0f-f385-4fad-9825-2824f47dd8db\") " pod="openstack/rabbitmq-server-0" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.768597 4806 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f18b8381-c1f9-4e48-9639-adfc5a406726-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:54 crc kubenswrapper[4806]: I1204 04:18:54.896879 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.316405 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.352401 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.360220 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.374390 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.395732 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.397327 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.401633 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.402136 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.402480 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.402752 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.405188 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.405442 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-52qcl" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.406650 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.413554 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.440882 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b7d7449-e532-4a32-bfcd-dc9f92767bd6" path="/var/lib/kubelet/pods/0b7d7449-e532-4a32-bfcd-dc9f92767bd6/volumes" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.442914 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f18b8381-c1f9-4e48-9639-adfc5a406726" path="/var/lib/kubelet/pods/f18b8381-c1f9-4e48-9639-adfc5a406726/volumes" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.587072 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/17d9cfc5-f7f9-442f-8174-2b61041b8f70-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.587213 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.587247 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/17d9cfc5-f7f9-442f-8174-2b61041b8f70-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.587275 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/17d9cfc5-f7f9-442f-8174-2b61041b8f70-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.587306 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/17d9cfc5-f7f9-442f-8174-2b61041b8f70-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.587342 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17d9cfc5-f7f9-442f-8174-2b61041b8f70-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.587362 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxvxj\" (UniqueName: \"kubernetes.io/projected/17d9cfc5-f7f9-442f-8174-2b61041b8f70-kube-api-access-lxvxj\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.587408 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/17d9cfc5-f7f9-442f-8174-2b61041b8f70-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.587477 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/17d9cfc5-f7f9-442f-8174-2b61041b8f70-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.587500 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/17d9cfc5-f7f9-442f-8174-2b61041b8f70-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.587532 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/17d9cfc5-f7f9-442f-8174-2b61041b8f70-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.689682 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/17d9cfc5-f7f9-442f-8174-2b61041b8f70-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.689741 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/17d9cfc5-f7f9-442f-8174-2b61041b8f70-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.689780 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/17d9cfc5-f7f9-442f-8174-2b61041b8f70-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.689809 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/17d9cfc5-f7f9-442f-8174-2b61041b8f70-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.689856 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.689871 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/17d9cfc5-f7f9-442f-8174-2b61041b8f70-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.689890 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/17d9cfc5-f7f9-442f-8174-2b61041b8f70-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.689920 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/17d9cfc5-f7f9-442f-8174-2b61041b8f70-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.689964 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17d9cfc5-f7f9-442f-8174-2b61041b8f70-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.689989 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxvxj\" (UniqueName: \"kubernetes.io/projected/17d9cfc5-f7f9-442f-8174-2b61041b8f70-kube-api-access-lxvxj\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.690025 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/17d9cfc5-f7f9-442f-8174-2b61041b8f70-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.690400 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/17d9cfc5-f7f9-442f-8174-2b61041b8f70-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.690524 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/17d9cfc5-f7f9-442f-8174-2b61041b8f70-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.690974 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/17d9cfc5-f7f9-442f-8174-2b61041b8f70-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.691286 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17d9cfc5-f7f9-442f-8174-2b61041b8f70-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.691511 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.691622 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/17d9cfc5-f7f9-442f-8174-2b61041b8f70-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.696414 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/17d9cfc5-f7f9-442f-8174-2b61041b8f70-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.699964 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/17d9cfc5-f7f9-442f-8174-2b61041b8f70-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.700008 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/17d9cfc5-f7f9-442f-8174-2b61041b8f70-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.704594 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/17d9cfc5-f7f9-442f-8174-2b61041b8f70-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.716792 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxvxj\" (UniqueName: \"kubernetes.io/projected/17d9cfc5-f7f9-442f-8174-2b61041b8f70-kube-api-access-lxvxj\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.731538 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"17d9cfc5-f7f9-442f-8174-2b61041b8f70\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:55 crc kubenswrapper[4806]: I1204 04:18:55.943992 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:18:56 crc kubenswrapper[4806]: I1204 04:18:56.324843 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5a2a4a0f-f385-4fad-9825-2824f47dd8db","Type":"ContainerStarted","Data":"34cedfa4606766629483aa10c02cacddf8d6fd90424771d92f7b4a9672fcb6c2"} Dec 04 04:18:56 crc kubenswrapper[4806]: I1204 04:18:56.415721 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 04:18:56 crc kubenswrapper[4806]: I1204 04:18:56.770563 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-bgrnm"] Dec 04 04:18:56 crc kubenswrapper[4806]: I1204 04:18:56.772280 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:56 crc kubenswrapper[4806]: I1204 04:18:56.777261 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 04 04:18:56 crc kubenswrapper[4806]: I1204 04:18:56.799447 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-bgrnm"] Dec 04 04:18:56 crc kubenswrapper[4806]: I1204 04:18:56.910205 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:56 crc kubenswrapper[4806]: I1204 04:18:56.910268 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:56 crc kubenswrapper[4806]: I1204 04:18:56.910295 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:56 crc kubenswrapper[4806]: I1204 04:18:56.910397 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:56 crc kubenswrapper[4806]: I1204 04:18:56.910702 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:56 crc kubenswrapper[4806]: I1204 04:18:56.910737 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79gzf\" (UniqueName: \"kubernetes.io/projected/598952f2-6d90-4da7-bce1-991d672074b2-kube-api-access-79gzf\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:56 crc kubenswrapper[4806]: I1204 04:18:56.910852 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-config\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:56 crc kubenswrapper[4806]: I1204 04:18:56.953654 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-bgrnm"] Dec 04 04:18:56 crc kubenswrapper[4806]: E1204 04:18:56.953992 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-79gzf openstack-edpm-ipam ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" podUID="598952f2-6d90-4da7-bce1-991d672074b2" Dec 04 04:18:56 crc kubenswrapper[4806]: I1204 04:18:56.996522 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b7d8cccff-7r9mb"] Dec 04 04:18:56 crc kubenswrapper[4806]: I1204 04:18:56.998063 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.013432 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.013488 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79gzf\" (UniqueName: \"kubernetes.io/projected/598952f2-6d90-4da7-bce1-991d672074b2-kube-api-access-79gzf\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.013529 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-config\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.013580 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.013601 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.013617 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.013647 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.014607 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.015124 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.015604 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.016144 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.016672 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.018998 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7d8cccff-7r9mb"] Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.022951 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-config\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.047940 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.047981 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.051214 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79gzf\" (UniqueName: \"kubernetes.io/projected/598952f2-6d90-4da7-bce1-991d672074b2-kube-api-access-79gzf\") pod \"dnsmasq-dns-79bd4cc8c9-bgrnm\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.115494 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-ovsdbserver-nb\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.115566 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-dns-svc\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.115613 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-config\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.115641 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-dns-swift-storage-0\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.115711 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-ovsdbserver-sb\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.115749 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7pg6\" (UniqueName: \"kubernetes.io/projected/aa0dbd11-2fbb-4960-8446-d60a18c6c776-kube-api-access-n7pg6\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.115782 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.217227 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-config\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.217341 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-dns-swift-storage-0\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.218231 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-ovsdbserver-sb\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.218348 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-dns-swift-storage-0\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.218359 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7pg6\" (UniqueName: \"kubernetes.io/projected/aa0dbd11-2fbb-4960-8446-d60a18c6c776-kube-api-access-n7pg6\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.218451 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.218520 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-ovsdbserver-nb\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.218561 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-dns-svc\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.219418 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-ovsdbserver-nb\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.219439 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.219442 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-dns-svc\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.219558 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-config\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.220678 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-ovsdbserver-sb\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.242610 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7pg6\" (UniqueName: \"kubernetes.io/projected/aa0dbd11-2fbb-4960-8446-d60a18c6c776-kube-api-access-n7pg6\") pod \"dnsmasq-dns-5b7d8cccff-7r9mb\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.312751 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.359133 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5a2a4a0f-f385-4fad-9825-2824f47dd8db","Type":"ContainerStarted","Data":"92c79cc5bcd1cda4e5578290057ccc8466b3baf56e148f1269f38d358d512c32"} Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.363223 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.364070 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"17d9cfc5-f7f9-442f-8174-2b61041b8f70","Type":"ContainerStarted","Data":"ac3ff4390ad3bfe83fb705529a8f036c89edaf59c5aa64f35c09601743342445"} Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.459957 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.524539 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-ovsdbserver-sb\") pod \"598952f2-6d90-4da7-bce1-991d672074b2\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.524659 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-dns-swift-storage-0\") pod \"598952f2-6d90-4da7-bce1-991d672074b2\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.524716 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79gzf\" (UniqueName: \"kubernetes.io/projected/598952f2-6d90-4da7-bce1-991d672074b2-kube-api-access-79gzf\") pod \"598952f2-6d90-4da7-bce1-991d672074b2\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.524740 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-ovsdbserver-nb\") pod \"598952f2-6d90-4da7-bce1-991d672074b2\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.524784 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-openstack-edpm-ipam\") pod \"598952f2-6d90-4da7-bce1-991d672074b2\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.524828 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-dns-svc\") pod \"598952f2-6d90-4da7-bce1-991d672074b2\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.524892 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-config\") pod \"598952f2-6d90-4da7-bce1-991d672074b2\" (UID: \"598952f2-6d90-4da7-bce1-991d672074b2\") " Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.525235 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "598952f2-6d90-4da7-bce1-991d672074b2" (UID: "598952f2-6d90-4da7-bce1-991d672074b2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.526001 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "598952f2-6d90-4da7-bce1-991d672074b2" (UID: "598952f2-6d90-4da7-bce1-991d672074b2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.526269 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "598952f2-6d90-4da7-bce1-991d672074b2" (UID: "598952f2-6d90-4da7-bce1-991d672074b2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.526805 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "598952f2-6d90-4da7-bce1-991d672074b2" (UID: "598952f2-6d90-4da7-bce1-991d672074b2"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.526868 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-config" (OuterVolumeSpecName: "config") pod "598952f2-6d90-4da7-bce1-991d672074b2" (UID: "598952f2-6d90-4da7-bce1-991d672074b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.526971 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "598952f2-6d90-4da7-bce1-991d672074b2" (UID: "598952f2-6d90-4da7-bce1-991d672074b2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.527375 4806 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.527629 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.527639 4806 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.527649 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.527657 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.527665 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/598952f2-6d90-4da7-bce1-991d672074b2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.542987 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/598952f2-6d90-4da7-bce1-991d672074b2-kube-api-access-79gzf" (OuterVolumeSpecName: "kube-api-access-79gzf") pod "598952f2-6d90-4da7-bce1-991d672074b2" (UID: "598952f2-6d90-4da7-bce1-991d672074b2"). InnerVolumeSpecName "kube-api-access-79gzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.629648 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79gzf\" (UniqueName: \"kubernetes.io/projected/598952f2-6d90-4da7-bce1-991d672074b2-kube-api-access-79gzf\") on node \"crc\" DevicePath \"\"" Dec 04 04:18:57 crc kubenswrapper[4806]: I1204 04:18:57.839896 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7d8cccff-7r9mb"] Dec 04 04:18:57 crc kubenswrapper[4806]: W1204 04:18:57.858674 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa0dbd11_2fbb_4960_8446_d60a18c6c776.slice/crio-6bfefd2048c3b4b67b9305f8fa315a39c3054c5dbc1b4e2be51db9f2d5800011 WatchSource:0}: Error finding container 6bfefd2048c3b4b67b9305f8fa315a39c3054c5dbc1b4e2be51db9f2d5800011: Status 404 returned error can't find the container with id 6bfefd2048c3b4b67b9305f8fa315a39c3054c5dbc1b4e2be51db9f2d5800011 Dec 04 04:18:58 crc kubenswrapper[4806]: I1204 04:18:58.373059 4806 generic.go:334] "Generic (PLEG): container finished" podID="aa0dbd11-2fbb-4960-8446-d60a18c6c776" containerID="2e6a94ac72e28e8ed436f6a0ed1f91be0e4798d4d5caae9c95969dfa6f5c38e3" exitCode=0 Dec 04 04:18:58 crc kubenswrapper[4806]: I1204 04:18:58.373400 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" event={"ID":"aa0dbd11-2fbb-4960-8446-d60a18c6c776","Type":"ContainerDied","Data":"2e6a94ac72e28e8ed436f6a0ed1f91be0e4798d4d5caae9c95969dfa6f5c38e3"} Dec 04 04:18:58 crc kubenswrapper[4806]: I1204 04:18:58.373430 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" event={"ID":"aa0dbd11-2fbb-4960-8446-d60a18c6c776","Type":"ContainerStarted","Data":"6bfefd2048c3b4b67b9305f8fa315a39c3054c5dbc1b4e2be51db9f2d5800011"} Dec 04 04:18:58 crc kubenswrapper[4806]: I1204 04:18:58.377992 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"17d9cfc5-f7f9-442f-8174-2b61041b8f70","Type":"ContainerStarted","Data":"23fa2923bfcc4edc218d41230cf9fe83ca301a2e9eb6cc21d9179843573c8c93"} Dec 04 04:18:58 crc kubenswrapper[4806]: I1204 04:18:58.378042 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-bgrnm" Dec 04 04:18:58 crc kubenswrapper[4806]: I1204 04:18:58.696879 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-bgrnm"] Dec 04 04:18:58 crc kubenswrapper[4806]: I1204 04:18:58.722382 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-bgrnm"] Dec 04 04:18:59 crc kubenswrapper[4806]: I1204 04:18:59.403980 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" event={"ID":"aa0dbd11-2fbb-4960-8446-d60a18c6c776","Type":"ContainerStarted","Data":"cc8516e3e42fc34588e518220726fcc7b1333b9f9a4bc5bf79ee92ae4f908035"} Dec 04 04:18:59 crc kubenswrapper[4806]: I1204 04:18:59.404052 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:18:59 crc kubenswrapper[4806]: I1204 04:18:59.431079 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" podStartSLOduration=3.431046833 podStartE2EDuration="3.431046833s" podCreationTimestamp="2025-12-04 04:18:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:18:59.420091087 +0000 UTC m=+1454.278604035" watchObservedRunningTime="2025-12-04 04:18:59.431046833 +0000 UTC m=+1454.289559761" Dec 04 04:18:59 crc kubenswrapper[4806]: I1204 04:18:59.435446 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="598952f2-6d90-4da7-bce1-991d672074b2" path="/var/lib/kubelet/pods/598952f2-6d90-4da7-bce1-991d672074b2/volumes" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.314765 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.374461 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-mp95p"] Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.374714 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" podUID="66a62f3e-3d15-4d90-b725-69a85adb938d" containerName="dnsmasq-dns" containerID="cri-o://23fada87b780f82825f808069deed276a683d5dca7a736af5a30fac5410a372b" gracePeriod=10 Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.541996 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d7677974f-lk2tx"] Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.543595 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.600014 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d7677974f-lk2tx"] Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.601396 4806 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" podUID="66a62f3e-3d15-4d90-b725-69a85adb938d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.197:5353: connect: connection refused" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.719913 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-dns-swift-storage-0\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.720100 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-openstack-edpm-ipam\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.720307 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dn5n\" (UniqueName: \"kubernetes.io/projected/5e42fed2-6410-480b-ad51-7c8a5c383eea-kube-api-access-9dn5n\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.720405 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-config\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.720503 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-ovsdbserver-sb\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.720555 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-ovsdbserver-nb\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.720694 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-dns-svc\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.821953 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-ovsdbserver-sb\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.822396 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-ovsdbserver-nb\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.822464 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-dns-svc\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.822497 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-dns-swift-storage-0\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.822525 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-openstack-edpm-ipam\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.822602 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dn5n\" (UniqueName: \"kubernetes.io/projected/5e42fed2-6410-480b-ad51-7c8a5c383eea-kube-api-access-9dn5n\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.822646 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-config\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.823311 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-ovsdbserver-sb\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.823535 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-dns-svc\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.824242 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-ovsdbserver-nb\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.825326 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-config\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.825748 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-dns-swift-storage-0\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.829235 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5e42fed2-6410-480b-ad51-7c8a5c383eea-openstack-edpm-ipam\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.850042 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dn5n\" (UniqueName: \"kubernetes.io/projected/5e42fed2-6410-480b-ad51-7c8a5c383eea-kube-api-access-9dn5n\") pod \"dnsmasq-dns-d7677974f-lk2tx\" (UID: \"5e42fed2-6410-480b-ad51-7c8a5c383eea\") " pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:07 crc kubenswrapper[4806]: I1204 04:19:07.886488 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.009154 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.128882 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-config\") pod \"66a62f3e-3d15-4d90-b725-69a85adb938d\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.128963 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf4fv\" (UniqueName: \"kubernetes.io/projected/66a62f3e-3d15-4d90-b725-69a85adb938d-kube-api-access-bf4fv\") pod \"66a62f3e-3d15-4d90-b725-69a85adb938d\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.128993 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-dns-svc\") pod \"66a62f3e-3d15-4d90-b725-69a85adb938d\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.129022 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-dns-swift-storage-0\") pod \"66a62f3e-3d15-4d90-b725-69a85adb938d\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.129106 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-ovsdbserver-sb\") pod \"66a62f3e-3d15-4d90-b725-69a85adb938d\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.129235 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-ovsdbserver-nb\") pod \"66a62f3e-3d15-4d90-b725-69a85adb938d\" (UID: \"66a62f3e-3d15-4d90-b725-69a85adb938d\") " Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.136142 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66a62f3e-3d15-4d90-b725-69a85adb938d-kube-api-access-bf4fv" (OuterVolumeSpecName: "kube-api-access-bf4fv") pod "66a62f3e-3d15-4d90-b725-69a85adb938d" (UID: "66a62f3e-3d15-4d90-b725-69a85adb938d"). InnerVolumeSpecName "kube-api-access-bf4fv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.200196 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "66a62f3e-3d15-4d90-b725-69a85adb938d" (UID: "66a62f3e-3d15-4d90-b725-69a85adb938d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.226371 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "66a62f3e-3d15-4d90-b725-69a85adb938d" (UID: "66a62f3e-3d15-4d90-b725-69a85adb938d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.226559 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "66a62f3e-3d15-4d90-b725-69a85adb938d" (UID: "66a62f3e-3d15-4d90-b725-69a85adb938d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.231245 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.231278 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf4fv\" (UniqueName: \"kubernetes.io/projected/66a62f3e-3d15-4d90-b725-69a85adb938d-kube-api-access-bf4fv\") on node \"crc\" DevicePath \"\"" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.231292 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.231301 4806 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.235099 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-config" (OuterVolumeSpecName: "config") pod "66a62f3e-3d15-4d90-b725-69a85adb938d" (UID: "66a62f3e-3d15-4d90-b725-69a85adb938d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.243693 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "66a62f3e-3d15-4d90-b725-69a85adb938d" (UID: "66a62f3e-3d15-4d90-b725-69a85adb938d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.332937 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.332973 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66a62f3e-3d15-4d90-b725-69a85adb938d-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.448962 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d7677974f-lk2tx"] Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.489723 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d7677974f-lk2tx" event={"ID":"5e42fed2-6410-480b-ad51-7c8a5c383eea","Type":"ContainerStarted","Data":"e2f3d7763c01b13a976b7e4a76d39a90c17069db8778ccab9fe18c48b7c5f538"} Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.495080 4806 generic.go:334] "Generic (PLEG): container finished" podID="66a62f3e-3d15-4d90-b725-69a85adb938d" containerID="23fada87b780f82825f808069deed276a683d5dca7a736af5a30fac5410a372b" exitCode=0 Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.495134 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" event={"ID":"66a62f3e-3d15-4d90-b725-69a85adb938d","Type":"ContainerDied","Data":"23fada87b780f82825f808069deed276a683d5dca7a736af5a30fac5410a372b"} Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.495289 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" event={"ID":"66a62f3e-3d15-4d90-b725-69a85adb938d","Type":"ContainerDied","Data":"5d39bb6776d16b002d3b100a2d20a8026b8d899701e1c9b04b40223037f38396"} Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.495314 4806 scope.go:117] "RemoveContainer" containerID="23fada87b780f82825f808069deed276a683d5dca7a736af5a30fac5410a372b" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.495467 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-mp95p" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.642268 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-mp95p"] Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.653417 4806 scope.go:117] "RemoveContainer" containerID="109170b5a39a76faf90a107fbdfb0b74637bb8c5547b7e55f73bd0c2a1685006" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.670084 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-mp95p"] Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.705890 4806 scope.go:117] "RemoveContainer" containerID="23fada87b780f82825f808069deed276a683d5dca7a736af5a30fac5410a372b" Dec 04 04:19:08 crc kubenswrapper[4806]: E1204 04:19:08.706319 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23fada87b780f82825f808069deed276a683d5dca7a736af5a30fac5410a372b\": container with ID starting with 23fada87b780f82825f808069deed276a683d5dca7a736af5a30fac5410a372b not found: ID does not exist" containerID="23fada87b780f82825f808069deed276a683d5dca7a736af5a30fac5410a372b" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.706349 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23fada87b780f82825f808069deed276a683d5dca7a736af5a30fac5410a372b"} err="failed to get container status \"23fada87b780f82825f808069deed276a683d5dca7a736af5a30fac5410a372b\": rpc error: code = NotFound desc = could not find container \"23fada87b780f82825f808069deed276a683d5dca7a736af5a30fac5410a372b\": container with ID starting with 23fada87b780f82825f808069deed276a683d5dca7a736af5a30fac5410a372b not found: ID does not exist" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.706369 4806 scope.go:117] "RemoveContainer" containerID="109170b5a39a76faf90a107fbdfb0b74637bb8c5547b7e55f73bd0c2a1685006" Dec 04 04:19:08 crc kubenswrapper[4806]: E1204 04:19:08.706595 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"109170b5a39a76faf90a107fbdfb0b74637bb8c5547b7e55f73bd0c2a1685006\": container with ID starting with 109170b5a39a76faf90a107fbdfb0b74637bb8c5547b7e55f73bd0c2a1685006 not found: ID does not exist" containerID="109170b5a39a76faf90a107fbdfb0b74637bb8c5547b7e55f73bd0c2a1685006" Dec 04 04:19:08 crc kubenswrapper[4806]: I1204 04:19:08.706620 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"109170b5a39a76faf90a107fbdfb0b74637bb8c5547b7e55f73bd0c2a1685006"} err="failed to get container status \"109170b5a39a76faf90a107fbdfb0b74637bb8c5547b7e55f73bd0c2a1685006\": rpc error: code = NotFound desc = could not find container \"109170b5a39a76faf90a107fbdfb0b74637bb8c5547b7e55f73bd0c2a1685006\": container with ID starting with 109170b5a39a76faf90a107fbdfb0b74637bb8c5547b7e55f73bd0c2a1685006 not found: ID does not exist" Dec 04 04:19:09 crc kubenswrapper[4806]: I1204 04:19:09.434851 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66a62f3e-3d15-4d90-b725-69a85adb938d" path="/var/lib/kubelet/pods/66a62f3e-3d15-4d90-b725-69a85adb938d/volumes" Dec 04 04:19:09 crc kubenswrapper[4806]: I1204 04:19:09.506867 4806 generic.go:334] "Generic (PLEG): container finished" podID="5e42fed2-6410-480b-ad51-7c8a5c383eea" containerID="9124de411b9809b089e0ef566717b0ece11f3fa2376fbd5deddd57cf06ee6abe" exitCode=0 Dec 04 04:19:09 crc kubenswrapper[4806]: I1204 04:19:09.506974 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d7677974f-lk2tx" event={"ID":"5e42fed2-6410-480b-ad51-7c8a5c383eea","Type":"ContainerDied","Data":"9124de411b9809b089e0ef566717b0ece11f3fa2376fbd5deddd57cf06ee6abe"} Dec 04 04:19:10 crc kubenswrapper[4806]: I1204 04:19:10.520899 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d7677974f-lk2tx" event={"ID":"5e42fed2-6410-480b-ad51-7c8a5c383eea","Type":"ContainerStarted","Data":"c86bb2be19a2128109183c9e545d521aaeaf5f6bcadb974ac76c9658efc05cf4"} Dec 04 04:19:10 crc kubenswrapper[4806]: I1204 04:19:10.521275 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:10 crc kubenswrapper[4806]: I1204 04:19:10.541512 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d7677974f-lk2tx" podStartSLOduration=3.541495749 podStartE2EDuration="3.541495749s" podCreationTimestamp="2025-12-04 04:19:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:19:10.537458452 +0000 UTC m=+1465.395971380" watchObservedRunningTime="2025-12-04 04:19:10.541495749 +0000 UTC m=+1465.400008667" Dec 04 04:19:17 crc kubenswrapper[4806]: I1204 04:19:17.888111 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d7677974f-lk2tx" Dec 04 04:19:17 crc kubenswrapper[4806]: I1204 04:19:17.948584 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7d8cccff-7r9mb"] Dec 04 04:19:17 crc kubenswrapper[4806]: I1204 04:19:17.948873 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" podUID="aa0dbd11-2fbb-4960-8446-d60a18c6c776" containerName="dnsmasq-dns" containerID="cri-o://cc8516e3e42fc34588e518220726fcc7b1333b9f9a4bc5bf79ee92ae4f908035" gracePeriod=10 Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.622663 4806 generic.go:334] "Generic (PLEG): container finished" podID="aa0dbd11-2fbb-4960-8446-d60a18c6c776" containerID="cc8516e3e42fc34588e518220726fcc7b1333b9f9a4bc5bf79ee92ae4f908035" exitCode=0 Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.623069 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" event={"ID":"aa0dbd11-2fbb-4960-8446-d60a18c6c776","Type":"ContainerDied","Data":"cc8516e3e42fc34588e518220726fcc7b1333b9f9a4bc5bf79ee92ae4f908035"} Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.623106 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" event={"ID":"aa0dbd11-2fbb-4960-8446-d60a18c6c776","Type":"ContainerDied","Data":"6bfefd2048c3b4b67b9305f8fa315a39c3054c5dbc1b4e2be51db9f2d5800011"} Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.623118 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bfefd2048c3b4b67b9305f8fa315a39c3054c5dbc1b4e2be51db9f2d5800011" Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.681737 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.839004 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-ovsdbserver-sb\") pod \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.839111 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-openstack-edpm-ipam\") pod \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.839187 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-config\") pod \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.839326 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7pg6\" (UniqueName: \"kubernetes.io/projected/aa0dbd11-2fbb-4960-8446-d60a18c6c776-kube-api-access-n7pg6\") pod \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.839381 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-dns-swift-storage-0\") pod \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.839409 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-ovsdbserver-nb\") pod \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.839442 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-dns-svc\") pod \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\" (UID: \"aa0dbd11-2fbb-4960-8446-d60a18c6c776\") " Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.845571 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa0dbd11-2fbb-4960-8446-d60a18c6c776-kube-api-access-n7pg6" (OuterVolumeSpecName: "kube-api-access-n7pg6") pod "aa0dbd11-2fbb-4960-8446-d60a18c6c776" (UID: "aa0dbd11-2fbb-4960-8446-d60a18c6c776"). InnerVolumeSpecName "kube-api-access-n7pg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.901538 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "aa0dbd11-2fbb-4960-8446-d60a18c6c776" (UID: "aa0dbd11-2fbb-4960-8446-d60a18c6c776"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.915952 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-config" (OuterVolumeSpecName: "config") pod "aa0dbd11-2fbb-4960-8446-d60a18c6c776" (UID: "aa0dbd11-2fbb-4960-8446-d60a18c6c776"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.916348 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aa0dbd11-2fbb-4960-8446-d60a18c6c776" (UID: "aa0dbd11-2fbb-4960-8446-d60a18c6c776"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.934204 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aa0dbd11-2fbb-4960-8446-d60a18c6c776" (UID: "aa0dbd11-2fbb-4960-8446-d60a18c6c776"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.937497 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aa0dbd11-2fbb-4960-8446-d60a18c6c776" (UID: "aa0dbd11-2fbb-4960-8446-d60a18c6c776"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.942073 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7pg6\" (UniqueName: \"kubernetes.io/projected/aa0dbd11-2fbb-4960-8446-d60a18c6c776-kube-api-access-n7pg6\") on node \"crc\" DevicePath \"\"" Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.942101 4806 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.942114 4806 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.942122 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.942130 4806 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.942138 4806 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-config\") on node \"crc\" DevicePath \"\"" Dec 04 04:19:18 crc kubenswrapper[4806]: I1204 04:19:18.943120 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aa0dbd11-2fbb-4960-8446-d60a18c6c776" (UID: "aa0dbd11-2fbb-4960-8446-d60a18c6c776"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:19:19 crc kubenswrapper[4806]: I1204 04:19:19.043988 4806 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa0dbd11-2fbb-4960-8446-d60a18c6c776-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 04:19:19 crc kubenswrapper[4806]: I1204 04:19:19.633006 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7d8cccff-7r9mb" Dec 04 04:19:19 crc kubenswrapper[4806]: I1204 04:19:19.653460 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7d8cccff-7r9mb"] Dec 04 04:19:19 crc kubenswrapper[4806]: I1204 04:19:19.667083 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b7d8cccff-7r9mb"] Dec 04 04:19:21 crc kubenswrapper[4806]: I1204 04:19:21.437295 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa0dbd11-2fbb-4960-8446-d60a18c6c776" path="/var/lib/kubelet/pods/aa0dbd11-2fbb-4960-8446-d60a18c6c776/volumes" Dec 04 04:19:27 crc kubenswrapper[4806]: I1204 04:19:27.047307 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:19:27 crc kubenswrapper[4806]: I1204 04:19:27.047754 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:19:27 crc kubenswrapper[4806]: I1204 04:19:27.047796 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 04:19:27 crc kubenswrapper[4806]: I1204 04:19:27.048331 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 04:19:27 crc kubenswrapper[4806]: I1204 04:19:27.048380 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" gracePeriod=600 Dec 04 04:19:27 crc kubenswrapper[4806]: E1204 04:19:27.179730 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:19:27 crc kubenswrapper[4806]: E1204 04:19:27.295571 4806 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0f9623b_44ce_43c2_b990_a75db2d17ff8.slice/crio-conmon-552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc.scope\": RecentStats: unable to find data in memory cache]" Dec 04 04:19:27 crc kubenswrapper[4806]: I1204 04:19:27.707037 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" exitCode=0 Dec 04 04:19:27 crc kubenswrapper[4806]: I1204 04:19:27.707081 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc"} Dec 04 04:19:27 crc kubenswrapper[4806]: I1204 04:19:27.707113 4806 scope.go:117] "RemoveContainer" containerID="95ac262c2400e25c649618e673d25b593b945bc15ad472113e55f250a2d5fdb0" Dec 04 04:19:27 crc kubenswrapper[4806]: I1204 04:19:27.707699 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:19:27 crc kubenswrapper[4806]: E1204 04:19:27.710882 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:19:29 crc kubenswrapper[4806]: I1204 04:19:29.728636 4806 generic.go:334] "Generic (PLEG): container finished" podID="5a2a4a0f-f385-4fad-9825-2824f47dd8db" containerID="92c79cc5bcd1cda4e5578290057ccc8466b3baf56e148f1269f38d358d512c32" exitCode=0 Dec 04 04:19:29 crc kubenswrapper[4806]: I1204 04:19:29.728732 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5a2a4a0f-f385-4fad-9825-2824f47dd8db","Type":"ContainerDied","Data":"92c79cc5bcd1cda4e5578290057ccc8466b3baf56e148f1269f38d358d512c32"} Dec 04 04:19:30 crc kubenswrapper[4806]: I1204 04:19:30.742381 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5a2a4a0f-f385-4fad-9825-2824f47dd8db","Type":"ContainerStarted","Data":"7f60bbc63c1edf1b3b7b996d77d6ad755f986bcf968c9aade7e18a9cb2e1680e"} Dec 04 04:19:30 crc kubenswrapper[4806]: I1204 04:19:30.743085 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 04 04:19:30 crc kubenswrapper[4806]: I1204 04:19:30.745437 4806 generic.go:334] "Generic (PLEG): container finished" podID="17d9cfc5-f7f9-442f-8174-2b61041b8f70" containerID="23fa2923bfcc4edc218d41230cf9fe83ca301a2e9eb6cc21d9179843573c8c93" exitCode=0 Dec 04 04:19:30 crc kubenswrapper[4806]: I1204 04:19:30.745480 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"17d9cfc5-f7f9-442f-8174-2b61041b8f70","Type":"ContainerDied","Data":"23fa2923bfcc4edc218d41230cf9fe83ca301a2e9eb6cc21d9179843573c8c93"} Dec 04 04:19:30 crc kubenswrapper[4806]: I1204 04:19:30.910330 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.910309784 podStartE2EDuration="36.910309784s" podCreationTimestamp="2025-12-04 04:18:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:19:30.873558401 +0000 UTC m=+1485.732071329" watchObservedRunningTime="2025-12-04 04:19:30.910309784 +0000 UTC m=+1485.768822712" Dec 04 04:19:31 crc kubenswrapper[4806]: I1204 04:19:31.757748 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"17d9cfc5-f7f9-442f-8174-2b61041b8f70","Type":"ContainerStarted","Data":"49f106060a198e1bb94a14e3b77f5b7c981792037ae2d04d031203c3b9a76385"} Dec 04 04:19:31 crc kubenswrapper[4806]: I1204 04:19:31.758648 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:19:31 crc kubenswrapper[4806]: I1204 04:19:31.790738 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.79071733 podStartE2EDuration="36.79071733s" podCreationTimestamp="2025-12-04 04:18:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 04:19:31.783384028 +0000 UTC m=+1486.641896966" watchObservedRunningTime="2025-12-04 04:19:31.79071733 +0000 UTC m=+1486.649230258" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.424717 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:19:41 crc kubenswrapper[4806]: E1204 04:19:41.426422 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.608777 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld"] Dec 04 04:19:41 crc kubenswrapper[4806]: E1204 04:19:41.609583 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa0dbd11-2fbb-4960-8446-d60a18c6c776" containerName="init" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.609683 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa0dbd11-2fbb-4960-8446-d60a18c6c776" containerName="init" Dec 04 04:19:41 crc kubenswrapper[4806]: E1204 04:19:41.609780 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa0dbd11-2fbb-4960-8446-d60a18c6c776" containerName="dnsmasq-dns" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.609860 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa0dbd11-2fbb-4960-8446-d60a18c6c776" containerName="dnsmasq-dns" Dec 04 04:19:41 crc kubenswrapper[4806]: E1204 04:19:41.609974 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66a62f3e-3d15-4d90-b725-69a85adb938d" containerName="init" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.610060 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="66a62f3e-3d15-4d90-b725-69a85adb938d" containerName="init" Dec 04 04:19:41 crc kubenswrapper[4806]: E1204 04:19:41.610154 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66a62f3e-3d15-4d90-b725-69a85adb938d" containerName="dnsmasq-dns" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.610242 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="66a62f3e-3d15-4d90-b725-69a85adb938d" containerName="dnsmasq-dns" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.610555 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="66a62f3e-3d15-4d90-b725-69a85adb938d" containerName="dnsmasq-dns" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.610637 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa0dbd11-2fbb-4960-8446-d60a18c6c776" containerName="dnsmasq-dns" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.611633 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.614043 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-q8mqk" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.615457 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.616209 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.616346 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.629521 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld"] Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.705789 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn8dg\" (UniqueName: \"kubernetes.io/projected/410d8040-e0b5-4869-bb16-41e5b59487d3-kube-api-access-nn8dg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld\" (UID: \"410d8040-e0b5-4869-bb16-41e5b59487d3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.706269 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld\" (UID: \"410d8040-e0b5-4869-bb16-41e5b59487d3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.706815 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld\" (UID: \"410d8040-e0b5-4869-bb16-41e5b59487d3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.706909 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld\" (UID: \"410d8040-e0b5-4869-bb16-41e5b59487d3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.810158 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld\" (UID: \"410d8040-e0b5-4869-bb16-41e5b59487d3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.810238 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld\" (UID: \"410d8040-e0b5-4869-bb16-41e5b59487d3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.810363 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn8dg\" (UniqueName: \"kubernetes.io/projected/410d8040-e0b5-4869-bb16-41e5b59487d3-kube-api-access-nn8dg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld\" (UID: \"410d8040-e0b5-4869-bb16-41e5b59487d3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.810473 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld\" (UID: \"410d8040-e0b5-4869-bb16-41e5b59487d3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.818056 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld\" (UID: \"410d8040-e0b5-4869-bb16-41e5b59487d3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.818122 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld\" (UID: \"410d8040-e0b5-4869-bb16-41e5b59487d3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.820036 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld\" (UID: \"410d8040-e0b5-4869-bb16-41e5b59487d3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.831150 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn8dg\" (UniqueName: \"kubernetes.io/projected/410d8040-e0b5-4869-bb16-41e5b59487d3-kube-api-access-nn8dg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld\" (UID: \"410d8040-e0b5-4869-bb16-41e5b59487d3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" Dec 04 04:19:41 crc kubenswrapper[4806]: I1204 04:19:41.932062 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" Dec 04 04:19:42 crc kubenswrapper[4806]: I1204 04:19:42.802626 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld"] Dec 04 04:19:42 crc kubenswrapper[4806]: I1204 04:19:42.860286 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" event={"ID":"410d8040-e0b5-4869-bb16-41e5b59487d3","Type":"ContainerStarted","Data":"0ff1deb9a5693ef333b64127ad436fedeb04408855c2ab34d0ed7fc9f7db1f8c"} Dec 04 04:19:44 crc kubenswrapper[4806]: I1204 04:19:44.900219 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 04 04:19:45 crc kubenswrapper[4806]: I1204 04:19:45.953328 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 04 04:19:52 crc kubenswrapper[4806]: I1204 04:19:52.636780 4806 scope.go:117] "RemoveContainer" containerID="fa63f5e0190e9d9dd4e1c9e13a341601eb4661dec32b87ddf56918276f9bad62" Dec 04 04:19:55 crc kubenswrapper[4806]: I1204 04:19:55.434364 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:19:55 crc kubenswrapper[4806]: E1204 04:19:55.435261 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:19:57 crc kubenswrapper[4806]: I1204 04:19:57.885789 4806 scope.go:117] "RemoveContainer" containerID="818f720223746c8f111996df6c232ef39f5651b1db629200e3810d1486707a6f" Dec 04 04:19:58 crc kubenswrapper[4806]: I1204 04:19:58.332195 4806 scope.go:117] "RemoveContainer" containerID="5e8e3cdd2cdc41f32edc8768a9ed26b1a0264603e7c50dbcdba6f2059cd35412" Dec 04 04:19:58 crc kubenswrapper[4806]: I1204 04:19:58.437668 4806 scope.go:117] "RemoveContainer" containerID="e59290c926c0f12d51a460b02e72a020b41534db66e0140ba00eddb2776df964" Dec 04 04:19:58 crc kubenswrapper[4806]: I1204 04:19:58.560738 4806 scope.go:117] "RemoveContainer" containerID="12c5f4cdfff6da63c40c019002d93c0627dc153a33cc9d065f4ac4c3b969019d" Dec 04 04:19:58 crc kubenswrapper[4806]: I1204 04:19:58.594305 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:19:59 crc kubenswrapper[4806]: I1204 04:19:59.128017 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" event={"ID":"410d8040-e0b5-4869-bb16-41e5b59487d3","Type":"ContainerStarted","Data":"94d9cef48d947a93fe7be2ef31e9426d0801811c1de1ad6779441c8578462f9c"} Dec 04 04:19:59 crc kubenswrapper[4806]: I1204 04:19:59.164692 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" podStartSLOduration=2.3858128450000002 podStartE2EDuration="18.164674483s" podCreationTimestamp="2025-12-04 04:19:41 +0000 UTC" firstStartedPulling="2025-12-04 04:19:42.811948618 +0000 UTC m=+1497.670461546" lastFinishedPulling="2025-12-04 04:19:58.590810256 +0000 UTC m=+1513.449323184" observedRunningTime="2025-12-04 04:19:59.149487187 +0000 UTC m=+1514.008000135" watchObservedRunningTime="2025-12-04 04:19:59.164674483 +0000 UTC m=+1514.023187411" Dec 04 04:20:00 crc kubenswrapper[4806]: I1204 04:20:00.946132 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qprw5"] Dec 04 04:20:00 crc kubenswrapper[4806]: I1204 04:20:00.949107 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:00 crc kubenswrapper[4806]: I1204 04:20:00.961271 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qprw5"] Dec 04 04:20:01 crc kubenswrapper[4806]: I1204 04:20:01.059264 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98plz\" (UniqueName: \"kubernetes.io/projected/ab9c72b7-546f-4dba-99be-f8d02ab56446-kube-api-access-98plz\") pod \"certified-operators-qprw5\" (UID: \"ab9c72b7-546f-4dba-99be-f8d02ab56446\") " pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:01 crc kubenswrapper[4806]: I1204 04:20:01.059984 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab9c72b7-546f-4dba-99be-f8d02ab56446-catalog-content\") pod \"certified-operators-qprw5\" (UID: \"ab9c72b7-546f-4dba-99be-f8d02ab56446\") " pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:01 crc kubenswrapper[4806]: I1204 04:20:01.060115 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab9c72b7-546f-4dba-99be-f8d02ab56446-utilities\") pod \"certified-operators-qprw5\" (UID: \"ab9c72b7-546f-4dba-99be-f8d02ab56446\") " pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:01 crc kubenswrapper[4806]: I1204 04:20:01.162194 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab9c72b7-546f-4dba-99be-f8d02ab56446-catalog-content\") pod \"certified-operators-qprw5\" (UID: \"ab9c72b7-546f-4dba-99be-f8d02ab56446\") " pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:01 crc kubenswrapper[4806]: I1204 04:20:01.162254 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab9c72b7-546f-4dba-99be-f8d02ab56446-utilities\") pod \"certified-operators-qprw5\" (UID: \"ab9c72b7-546f-4dba-99be-f8d02ab56446\") " pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:01 crc kubenswrapper[4806]: I1204 04:20:01.162305 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98plz\" (UniqueName: \"kubernetes.io/projected/ab9c72b7-546f-4dba-99be-f8d02ab56446-kube-api-access-98plz\") pod \"certified-operators-qprw5\" (UID: \"ab9c72b7-546f-4dba-99be-f8d02ab56446\") " pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:01 crc kubenswrapper[4806]: I1204 04:20:01.162763 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab9c72b7-546f-4dba-99be-f8d02ab56446-catalog-content\") pod \"certified-operators-qprw5\" (UID: \"ab9c72b7-546f-4dba-99be-f8d02ab56446\") " pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:01 crc kubenswrapper[4806]: I1204 04:20:01.163088 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab9c72b7-546f-4dba-99be-f8d02ab56446-utilities\") pod \"certified-operators-qprw5\" (UID: \"ab9c72b7-546f-4dba-99be-f8d02ab56446\") " pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:01 crc kubenswrapper[4806]: I1204 04:20:01.183105 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98plz\" (UniqueName: \"kubernetes.io/projected/ab9c72b7-546f-4dba-99be-f8d02ab56446-kube-api-access-98plz\") pod \"certified-operators-qprw5\" (UID: \"ab9c72b7-546f-4dba-99be-f8d02ab56446\") " pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:01 crc kubenswrapper[4806]: I1204 04:20:01.293023 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:01 crc kubenswrapper[4806]: I1204 04:20:01.982883 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qprw5"] Dec 04 04:20:02 crc kubenswrapper[4806]: I1204 04:20:02.155077 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qprw5" event={"ID":"ab9c72b7-546f-4dba-99be-f8d02ab56446","Type":"ContainerStarted","Data":"627c5b6b23bcdde562042f951b3aa6638213017134c3922b97d9c0520b70851e"} Dec 04 04:20:03 crc kubenswrapper[4806]: I1204 04:20:03.166063 4806 generic.go:334] "Generic (PLEG): container finished" podID="ab9c72b7-546f-4dba-99be-f8d02ab56446" containerID="c6f292a1913f71d51260da3854e22fad41dbb9b946e5ac54f4e25058c5ed0418" exitCode=0 Dec 04 04:20:03 crc kubenswrapper[4806]: I1204 04:20:03.166124 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qprw5" event={"ID":"ab9c72b7-546f-4dba-99be-f8d02ab56446","Type":"ContainerDied","Data":"c6f292a1913f71d51260da3854e22fad41dbb9b946e5ac54f4e25058c5ed0418"} Dec 04 04:20:04 crc kubenswrapper[4806]: I1204 04:20:04.178510 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qprw5" event={"ID":"ab9c72b7-546f-4dba-99be-f8d02ab56446","Type":"ContainerStarted","Data":"c5a1942d115d7ea70511df2c56549cb94b7c71415f70793b9daf746434f49dba"} Dec 04 04:20:07 crc kubenswrapper[4806]: I1204 04:20:07.203993 4806 generic.go:334] "Generic (PLEG): container finished" podID="ab9c72b7-546f-4dba-99be-f8d02ab56446" containerID="c5a1942d115d7ea70511df2c56549cb94b7c71415f70793b9daf746434f49dba" exitCode=0 Dec 04 04:20:07 crc kubenswrapper[4806]: I1204 04:20:07.204326 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qprw5" event={"ID":"ab9c72b7-546f-4dba-99be-f8d02ab56446","Type":"ContainerDied","Data":"c5a1942d115d7ea70511df2c56549cb94b7c71415f70793b9daf746434f49dba"} Dec 04 04:20:08 crc kubenswrapper[4806]: I1204 04:20:08.217600 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qprw5" event={"ID":"ab9c72b7-546f-4dba-99be-f8d02ab56446","Type":"ContainerStarted","Data":"4c05c2bb870a47d0dfd8466b0646aaecb4bb05af29b3a0161fa9f81cc6a3a44d"} Dec 04 04:20:08 crc kubenswrapper[4806]: I1204 04:20:08.241494 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qprw5" podStartSLOduration=3.780249716 podStartE2EDuration="8.241469558s" podCreationTimestamp="2025-12-04 04:20:00 +0000 UTC" firstStartedPulling="2025-12-04 04:20:03.168971473 +0000 UTC m=+1518.027484391" lastFinishedPulling="2025-12-04 04:20:07.630191285 +0000 UTC m=+1522.488704233" observedRunningTime="2025-12-04 04:20:08.237536944 +0000 UTC m=+1523.096049882" watchObservedRunningTime="2025-12-04 04:20:08.241469558 +0000 UTC m=+1523.099982496" Dec 04 04:20:11 crc kubenswrapper[4806]: I1204 04:20:11.267689 4806 generic.go:334] "Generic (PLEG): container finished" podID="410d8040-e0b5-4869-bb16-41e5b59487d3" containerID="94d9cef48d947a93fe7be2ef31e9426d0801811c1de1ad6779441c8578462f9c" exitCode=0 Dec 04 04:20:11 crc kubenswrapper[4806]: I1204 04:20:11.267725 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" event={"ID":"410d8040-e0b5-4869-bb16-41e5b59487d3","Type":"ContainerDied","Data":"94d9cef48d947a93fe7be2ef31e9426d0801811c1de1ad6779441c8578462f9c"} Dec 04 04:20:11 crc kubenswrapper[4806]: I1204 04:20:11.293602 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:11 crc kubenswrapper[4806]: I1204 04:20:11.294649 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:11 crc kubenswrapper[4806]: I1204 04:20:11.361387 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:11 crc kubenswrapper[4806]: I1204 04:20:11.423656 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:20:11 crc kubenswrapper[4806]: E1204 04:20:11.423997 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:20:12 crc kubenswrapper[4806]: I1204 04:20:12.338761 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:12 crc kubenswrapper[4806]: I1204 04:20:12.388795 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qprw5"] Dec 04 04:20:12 crc kubenswrapper[4806]: I1204 04:20:12.765680 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" Dec 04 04:20:12 crc kubenswrapper[4806]: I1204 04:20:12.940615 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nn8dg\" (UniqueName: \"kubernetes.io/projected/410d8040-e0b5-4869-bb16-41e5b59487d3-kube-api-access-nn8dg\") pod \"410d8040-e0b5-4869-bb16-41e5b59487d3\" (UID: \"410d8040-e0b5-4869-bb16-41e5b59487d3\") " Dec 04 04:20:12 crc kubenswrapper[4806]: I1204 04:20:12.940938 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-repo-setup-combined-ca-bundle\") pod \"410d8040-e0b5-4869-bb16-41e5b59487d3\" (UID: \"410d8040-e0b5-4869-bb16-41e5b59487d3\") " Dec 04 04:20:12 crc kubenswrapper[4806]: I1204 04:20:12.941022 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-ssh-key\") pod \"410d8040-e0b5-4869-bb16-41e5b59487d3\" (UID: \"410d8040-e0b5-4869-bb16-41e5b59487d3\") " Dec 04 04:20:12 crc kubenswrapper[4806]: I1204 04:20:12.941087 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-inventory\") pod \"410d8040-e0b5-4869-bb16-41e5b59487d3\" (UID: \"410d8040-e0b5-4869-bb16-41e5b59487d3\") " Dec 04 04:20:12 crc kubenswrapper[4806]: I1204 04:20:12.946986 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "410d8040-e0b5-4869-bb16-41e5b59487d3" (UID: "410d8040-e0b5-4869-bb16-41e5b59487d3"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:20:12 crc kubenswrapper[4806]: I1204 04:20:12.949029 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/410d8040-e0b5-4869-bb16-41e5b59487d3-kube-api-access-nn8dg" (OuterVolumeSpecName: "kube-api-access-nn8dg") pod "410d8040-e0b5-4869-bb16-41e5b59487d3" (UID: "410d8040-e0b5-4869-bb16-41e5b59487d3"). InnerVolumeSpecName "kube-api-access-nn8dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:20:12 crc kubenswrapper[4806]: I1204 04:20:12.982348 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "410d8040-e0b5-4869-bb16-41e5b59487d3" (UID: "410d8040-e0b5-4869-bb16-41e5b59487d3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:20:12 crc kubenswrapper[4806]: I1204 04:20:12.990910 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-inventory" (OuterVolumeSpecName: "inventory") pod "410d8040-e0b5-4869-bb16-41e5b59487d3" (UID: "410d8040-e0b5-4869-bb16-41e5b59487d3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.043466 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.043515 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nn8dg\" (UniqueName: \"kubernetes.io/projected/410d8040-e0b5-4869-bb16-41e5b59487d3-kube-api-access-nn8dg\") on node \"crc\" DevicePath \"\"" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.043533 4806 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.043543 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/410d8040-e0b5-4869-bb16-41e5b59487d3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.287126 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.287126 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld" event={"ID":"410d8040-e0b5-4869-bb16-41e5b59487d3","Type":"ContainerDied","Data":"0ff1deb9a5693ef333b64127ad436fedeb04408855c2ab34d0ed7fc9f7db1f8c"} Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.287186 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ff1deb9a5693ef333b64127ad436fedeb04408855c2ab34d0ed7fc9f7db1f8c" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.380017 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn"] Dec 04 04:20:13 crc kubenswrapper[4806]: E1204 04:20:13.380559 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410d8040-e0b5-4869-bb16-41e5b59487d3" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.380592 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="410d8040-e0b5-4869-bb16-41e5b59487d3" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.380828 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="410d8040-e0b5-4869-bb16-41e5b59487d3" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.382012 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.384450 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-q8mqk" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.384779 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.386182 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.386826 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.398670 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn"] Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.553685 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrftb\" (UniqueName: \"kubernetes.io/projected/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-kube-api-access-qrftb\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ds7xn\" (UID: \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.553834 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ds7xn\" (UID: \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.554004 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ds7xn\" (UID: \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.655704 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ds7xn\" (UID: \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.655797 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrftb\" (UniqueName: \"kubernetes.io/projected/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-kube-api-access-qrftb\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ds7xn\" (UID: \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.655855 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ds7xn\" (UID: \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.659597 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ds7xn\" (UID: \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.665451 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ds7xn\" (UID: \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.674944 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrftb\" (UniqueName: \"kubernetes.io/projected/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-kube-api-access-qrftb\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ds7xn\" (UID: \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" Dec 04 04:20:13 crc kubenswrapper[4806]: I1204 04:20:13.699842 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" Dec 04 04:20:14 crc kubenswrapper[4806]: I1204 04:20:14.296206 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qprw5" podUID="ab9c72b7-546f-4dba-99be-f8d02ab56446" containerName="registry-server" containerID="cri-o://4c05c2bb870a47d0dfd8466b0646aaecb4bb05af29b3a0161fa9f81cc6a3a44d" gracePeriod=2 Dec 04 04:20:14 crc kubenswrapper[4806]: I1204 04:20:14.322885 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn"] Dec 04 04:20:14 crc kubenswrapper[4806]: I1204 04:20:14.774050 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:14 crc kubenswrapper[4806]: I1204 04:20:14.886848 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab9c72b7-546f-4dba-99be-f8d02ab56446-catalog-content\") pod \"ab9c72b7-546f-4dba-99be-f8d02ab56446\" (UID: \"ab9c72b7-546f-4dba-99be-f8d02ab56446\") " Dec 04 04:20:14 crc kubenswrapper[4806]: I1204 04:20:14.887069 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98plz\" (UniqueName: \"kubernetes.io/projected/ab9c72b7-546f-4dba-99be-f8d02ab56446-kube-api-access-98plz\") pod \"ab9c72b7-546f-4dba-99be-f8d02ab56446\" (UID: \"ab9c72b7-546f-4dba-99be-f8d02ab56446\") " Dec 04 04:20:14 crc kubenswrapper[4806]: I1204 04:20:14.887099 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab9c72b7-546f-4dba-99be-f8d02ab56446-utilities\") pod \"ab9c72b7-546f-4dba-99be-f8d02ab56446\" (UID: \"ab9c72b7-546f-4dba-99be-f8d02ab56446\") " Dec 04 04:20:14 crc kubenswrapper[4806]: I1204 04:20:14.888741 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab9c72b7-546f-4dba-99be-f8d02ab56446-utilities" (OuterVolumeSpecName: "utilities") pod "ab9c72b7-546f-4dba-99be-f8d02ab56446" (UID: "ab9c72b7-546f-4dba-99be-f8d02ab56446"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:20:14 crc kubenswrapper[4806]: I1204 04:20:14.894790 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab9c72b7-546f-4dba-99be-f8d02ab56446-kube-api-access-98plz" (OuterVolumeSpecName: "kube-api-access-98plz") pod "ab9c72b7-546f-4dba-99be-f8d02ab56446" (UID: "ab9c72b7-546f-4dba-99be-f8d02ab56446"). InnerVolumeSpecName "kube-api-access-98plz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:20:14 crc kubenswrapper[4806]: I1204 04:20:14.969254 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab9c72b7-546f-4dba-99be-f8d02ab56446-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ab9c72b7-546f-4dba-99be-f8d02ab56446" (UID: "ab9c72b7-546f-4dba-99be-f8d02ab56446"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:20:14 crc kubenswrapper[4806]: I1204 04:20:14.993817 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab9c72b7-546f-4dba-99be-f8d02ab56446-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:20:14 crc kubenswrapper[4806]: I1204 04:20:14.993858 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98plz\" (UniqueName: \"kubernetes.io/projected/ab9c72b7-546f-4dba-99be-f8d02ab56446-kube-api-access-98plz\") on node \"crc\" DevicePath \"\"" Dec 04 04:20:14 crc kubenswrapper[4806]: I1204 04:20:14.993869 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab9c72b7-546f-4dba-99be-f8d02ab56446-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.305470 4806 generic.go:334] "Generic (PLEG): container finished" podID="ab9c72b7-546f-4dba-99be-f8d02ab56446" containerID="4c05c2bb870a47d0dfd8466b0646aaecb4bb05af29b3a0161fa9f81cc6a3a44d" exitCode=0 Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.305517 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qprw5" Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.305551 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qprw5" event={"ID":"ab9c72b7-546f-4dba-99be-f8d02ab56446","Type":"ContainerDied","Data":"4c05c2bb870a47d0dfd8466b0646aaecb4bb05af29b3a0161fa9f81cc6a3a44d"} Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.306024 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qprw5" event={"ID":"ab9c72b7-546f-4dba-99be-f8d02ab56446","Type":"ContainerDied","Data":"627c5b6b23bcdde562042f951b3aa6638213017134c3922b97d9c0520b70851e"} Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.306062 4806 scope.go:117] "RemoveContainer" containerID="4c05c2bb870a47d0dfd8466b0646aaecb4bb05af29b3a0161fa9f81cc6a3a44d" Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.308246 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" event={"ID":"ef28af41-0fb7-4e9a-adbe-c08fdf320e95","Type":"ContainerStarted","Data":"a50d6af92502463edaabb1ce78b9ac8359081dc984bbcc05d2bd5ca327ad2c3b"} Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.308279 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" event={"ID":"ef28af41-0fb7-4e9a-adbe-c08fdf320e95","Type":"ContainerStarted","Data":"d29b1e7c486cb6f257d6cfb7290b75cc6104e44628edb8972f994a111ce48b4b"} Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.332643 4806 scope.go:117] "RemoveContainer" containerID="c5a1942d115d7ea70511df2c56549cb94b7c71415f70793b9daf746434f49dba" Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.360518 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" podStartSLOduration=2.178626858 podStartE2EDuration="2.360496351s" podCreationTimestamp="2025-12-04 04:20:13 +0000 UTC" firstStartedPulling="2025-12-04 04:20:14.342605656 +0000 UTC m=+1529.201118584" lastFinishedPulling="2025-12-04 04:20:14.524475149 +0000 UTC m=+1529.382988077" observedRunningTime="2025-12-04 04:20:15.339571964 +0000 UTC m=+1530.198084892" watchObservedRunningTime="2025-12-04 04:20:15.360496351 +0000 UTC m=+1530.219009279" Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.363264 4806 scope.go:117] "RemoveContainer" containerID="c6f292a1913f71d51260da3854e22fad41dbb9b946e5ac54f4e25058c5ed0418" Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.371983 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qprw5"] Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.379499 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qprw5"] Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.412684 4806 scope.go:117] "RemoveContainer" containerID="4c05c2bb870a47d0dfd8466b0646aaecb4bb05af29b3a0161fa9f81cc6a3a44d" Dec 04 04:20:15 crc kubenswrapper[4806]: E1204 04:20:15.413444 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c05c2bb870a47d0dfd8466b0646aaecb4bb05af29b3a0161fa9f81cc6a3a44d\": container with ID starting with 4c05c2bb870a47d0dfd8466b0646aaecb4bb05af29b3a0161fa9f81cc6a3a44d not found: ID does not exist" containerID="4c05c2bb870a47d0dfd8466b0646aaecb4bb05af29b3a0161fa9f81cc6a3a44d" Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.413572 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c05c2bb870a47d0dfd8466b0646aaecb4bb05af29b3a0161fa9f81cc6a3a44d"} err="failed to get container status \"4c05c2bb870a47d0dfd8466b0646aaecb4bb05af29b3a0161fa9f81cc6a3a44d\": rpc error: code = NotFound desc = could not find container \"4c05c2bb870a47d0dfd8466b0646aaecb4bb05af29b3a0161fa9f81cc6a3a44d\": container with ID starting with 4c05c2bb870a47d0dfd8466b0646aaecb4bb05af29b3a0161fa9f81cc6a3a44d not found: ID does not exist" Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.413664 4806 scope.go:117] "RemoveContainer" containerID="c5a1942d115d7ea70511df2c56549cb94b7c71415f70793b9daf746434f49dba" Dec 04 04:20:15 crc kubenswrapper[4806]: E1204 04:20:15.413994 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5a1942d115d7ea70511df2c56549cb94b7c71415f70793b9daf746434f49dba\": container with ID starting with c5a1942d115d7ea70511df2c56549cb94b7c71415f70793b9daf746434f49dba not found: ID does not exist" containerID="c5a1942d115d7ea70511df2c56549cb94b7c71415f70793b9daf746434f49dba" Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.414022 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5a1942d115d7ea70511df2c56549cb94b7c71415f70793b9daf746434f49dba"} err="failed to get container status \"c5a1942d115d7ea70511df2c56549cb94b7c71415f70793b9daf746434f49dba\": rpc error: code = NotFound desc = could not find container \"c5a1942d115d7ea70511df2c56549cb94b7c71415f70793b9daf746434f49dba\": container with ID starting with c5a1942d115d7ea70511df2c56549cb94b7c71415f70793b9daf746434f49dba not found: ID does not exist" Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.414045 4806 scope.go:117] "RemoveContainer" containerID="c6f292a1913f71d51260da3854e22fad41dbb9b946e5ac54f4e25058c5ed0418" Dec 04 04:20:15 crc kubenswrapper[4806]: E1204 04:20:15.414268 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6f292a1913f71d51260da3854e22fad41dbb9b946e5ac54f4e25058c5ed0418\": container with ID starting with c6f292a1913f71d51260da3854e22fad41dbb9b946e5ac54f4e25058c5ed0418 not found: ID does not exist" containerID="c6f292a1913f71d51260da3854e22fad41dbb9b946e5ac54f4e25058c5ed0418" Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.414367 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6f292a1913f71d51260da3854e22fad41dbb9b946e5ac54f4e25058c5ed0418"} err="failed to get container status \"c6f292a1913f71d51260da3854e22fad41dbb9b946e5ac54f4e25058c5ed0418\": rpc error: code = NotFound desc = could not find container \"c6f292a1913f71d51260da3854e22fad41dbb9b946e5ac54f4e25058c5ed0418\": container with ID starting with c6f292a1913f71d51260da3854e22fad41dbb9b946e5ac54f4e25058c5ed0418 not found: ID does not exist" Dec 04 04:20:15 crc kubenswrapper[4806]: I1204 04:20:15.459432 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab9c72b7-546f-4dba-99be-f8d02ab56446" path="/var/lib/kubelet/pods/ab9c72b7-546f-4dba-99be-f8d02ab56446/volumes" Dec 04 04:20:18 crc kubenswrapper[4806]: I1204 04:20:18.341266 4806 generic.go:334] "Generic (PLEG): container finished" podID="ef28af41-0fb7-4e9a-adbe-c08fdf320e95" containerID="a50d6af92502463edaabb1ce78b9ac8359081dc984bbcc05d2bd5ca327ad2c3b" exitCode=0 Dec 04 04:20:18 crc kubenswrapper[4806]: I1204 04:20:18.341381 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" event={"ID":"ef28af41-0fb7-4e9a-adbe-c08fdf320e95","Type":"ContainerDied","Data":"a50d6af92502463edaabb1ce78b9ac8359081dc984bbcc05d2bd5ca327ad2c3b"} Dec 04 04:20:19 crc kubenswrapper[4806]: I1204 04:20:19.792360 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" Dec 04 04:20:19 crc kubenswrapper[4806]: I1204 04:20:19.917560 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-inventory\") pod \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\" (UID: \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\") " Dec 04 04:20:19 crc kubenswrapper[4806]: I1204 04:20:19.917775 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-ssh-key\") pod \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\" (UID: \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\") " Dec 04 04:20:19 crc kubenswrapper[4806]: I1204 04:20:19.917846 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrftb\" (UniqueName: \"kubernetes.io/projected/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-kube-api-access-qrftb\") pod \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\" (UID: \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\") " Dec 04 04:20:19 crc kubenswrapper[4806]: I1204 04:20:19.924995 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-kube-api-access-qrftb" (OuterVolumeSpecName: "kube-api-access-qrftb") pod "ef28af41-0fb7-4e9a-adbe-c08fdf320e95" (UID: "ef28af41-0fb7-4e9a-adbe-c08fdf320e95"). InnerVolumeSpecName "kube-api-access-qrftb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:20:19 crc kubenswrapper[4806]: E1204 04:20:19.946369 4806 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-inventory podName:ef28af41-0fb7-4e9a-adbe-c08fdf320e95 nodeName:}" failed. No retries permitted until 2025-12-04 04:20:20.446325888 +0000 UTC m=+1535.304838816 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-inventory") pod "ef28af41-0fb7-4e9a-adbe-c08fdf320e95" (UID: "ef28af41-0fb7-4e9a-adbe-c08fdf320e95") : error deleting /var/lib/kubelet/pods/ef28af41-0fb7-4e9a-adbe-c08fdf320e95/volume-subpaths: remove /var/lib/kubelet/pods/ef28af41-0fb7-4e9a-adbe-c08fdf320e95/volume-subpaths: no such file or directory Dec 04 04:20:19 crc kubenswrapper[4806]: I1204 04:20:19.951135 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ef28af41-0fb7-4e9a-adbe-c08fdf320e95" (UID: "ef28af41-0fb7-4e9a-adbe-c08fdf320e95"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.019840 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.019884 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrftb\" (UniqueName: \"kubernetes.io/projected/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-kube-api-access-qrftb\") on node \"crc\" DevicePath \"\"" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.360348 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" event={"ID":"ef28af41-0fb7-4e9a-adbe-c08fdf320e95","Type":"ContainerDied","Data":"d29b1e7c486cb6f257d6cfb7290b75cc6104e44628edb8972f994a111ce48b4b"} Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.360668 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d29b1e7c486cb6f257d6cfb7290b75cc6104e44628edb8972f994a111ce48b4b" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.360409 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ds7xn" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.445934 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8"] Dec 04 04:20:20 crc kubenswrapper[4806]: E1204 04:20:20.446383 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab9c72b7-546f-4dba-99be-f8d02ab56446" containerName="registry-server" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.446402 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab9c72b7-546f-4dba-99be-f8d02ab56446" containerName="registry-server" Dec 04 04:20:20 crc kubenswrapper[4806]: E1204 04:20:20.446416 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab9c72b7-546f-4dba-99be-f8d02ab56446" containerName="extract-content" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.446423 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab9c72b7-546f-4dba-99be-f8d02ab56446" containerName="extract-content" Dec 04 04:20:20 crc kubenswrapper[4806]: E1204 04:20:20.446439 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef28af41-0fb7-4e9a-adbe-c08fdf320e95" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.446449 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef28af41-0fb7-4e9a-adbe-c08fdf320e95" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 04 04:20:20 crc kubenswrapper[4806]: E1204 04:20:20.446467 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab9c72b7-546f-4dba-99be-f8d02ab56446" containerName="extract-utilities" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.446475 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab9c72b7-546f-4dba-99be-f8d02ab56446" containerName="extract-utilities" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.446702 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef28af41-0fb7-4e9a-adbe-c08fdf320e95" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.446737 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab9c72b7-546f-4dba-99be-f8d02ab56446" containerName="registry-server" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.447435 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.465658 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8"] Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.528074 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-inventory\") pod \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\" (UID: \"ef28af41-0fb7-4e9a-adbe-c08fdf320e95\") " Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.532757 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-inventory" (OuterVolumeSpecName: "inventory") pod "ef28af41-0fb7-4e9a-adbe-c08fdf320e95" (UID: "ef28af41-0fb7-4e9a-adbe-c08fdf320e95"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.630115 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8\" (UID: \"1787b37a-3067-4912-a658-4a8fd53faab6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.630271 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8\" (UID: \"1787b37a-3067-4912-a658-4a8fd53faab6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.630373 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8\" (UID: \"1787b37a-3067-4912-a658-4a8fd53faab6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.630504 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd4r9\" (UniqueName: \"kubernetes.io/projected/1787b37a-3067-4912-a658-4a8fd53faab6-kube-api-access-hd4r9\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8\" (UID: \"1787b37a-3067-4912-a658-4a8fd53faab6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.630570 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef28af41-0fb7-4e9a-adbe-c08fdf320e95-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.732891 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8\" (UID: \"1787b37a-3067-4912-a658-4a8fd53faab6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.733006 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8\" (UID: \"1787b37a-3067-4912-a658-4a8fd53faab6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.733100 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8\" (UID: \"1787b37a-3067-4912-a658-4a8fd53faab6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.733149 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd4r9\" (UniqueName: \"kubernetes.io/projected/1787b37a-3067-4912-a658-4a8fd53faab6-kube-api-access-hd4r9\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8\" (UID: \"1787b37a-3067-4912-a658-4a8fd53faab6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.737073 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8\" (UID: \"1787b37a-3067-4912-a658-4a8fd53faab6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.737090 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8\" (UID: \"1787b37a-3067-4912-a658-4a8fd53faab6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.738991 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8\" (UID: \"1787b37a-3067-4912-a658-4a8fd53faab6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.754620 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd4r9\" (UniqueName: \"kubernetes.io/projected/1787b37a-3067-4912-a658-4a8fd53faab6-kube-api-access-hd4r9\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8\" (UID: \"1787b37a-3067-4912-a658-4a8fd53faab6\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" Dec 04 04:20:20 crc kubenswrapper[4806]: I1204 04:20:20.764401 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" Dec 04 04:20:21 crc kubenswrapper[4806]: I1204 04:20:21.378455 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8"] Dec 04 04:20:21 crc kubenswrapper[4806]: W1204 04:20:21.383693 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1787b37a_3067_4912_a658_4a8fd53faab6.slice/crio-b2d7947e5bb91eeb123dfdbd555cc1ede902a15414fbdd68018c9df3a131c789 WatchSource:0}: Error finding container b2d7947e5bb91eeb123dfdbd555cc1ede902a15414fbdd68018c9df3a131c789: Status 404 returned error can't find the container with id b2d7947e5bb91eeb123dfdbd555cc1ede902a15414fbdd68018c9df3a131c789 Dec 04 04:20:22 crc kubenswrapper[4806]: I1204 04:20:22.379599 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" event={"ID":"1787b37a-3067-4912-a658-4a8fd53faab6","Type":"ContainerStarted","Data":"bf9ed1e2f3a08af0018cf98babc8fa1772a06f8599d234b8a0f4c6bc97bc3e32"} Dec 04 04:20:22 crc kubenswrapper[4806]: I1204 04:20:22.380154 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" event={"ID":"1787b37a-3067-4912-a658-4a8fd53faab6","Type":"ContainerStarted","Data":"b2d7947e5bb91eeb123dfdbd555cc1ede902a15414fbdd68018c9df3a131c789"} Dec 04 04:20:22 crc kubenswrapper[4806]: I1204 04:20:22.402820 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" podStartSLOduration=2.224797594 podStartE2EDuration="2.402801775s" podCreationTimestamp="2025-12-04 04:20:20 +0000 UTC" firstStartedPulling="2025-12-04 04:20:21.388624606 +0000 UTC m=+1536.247137534" lastFinishedPulling="2025-12-04 04:20:21.566628787 +0000 UTC m=+1536.425141715" observedRunningTime="2025-12-04 04:20:22.394633919 +0000 UTC m=+1537.253146847" watchObservedRunningTime="2025-12-04 04:20:22.402801775 +0000 UTC m=+1537.261314703" Dec 04 04:20:26 crc kubenswrapper[4806]: I1204 04:20:26.424413 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:20:26 crc kubenswrapper[4806]: E1204 04:20:26.426207 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:20:38 crc kubenswrapper[4806]: I1204 04:20:38.424320 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:20:38 crc kubenswrapper[4806]: E1204 04:20:38.425165 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:20:50 crc kubenswrapper[4806]: I1204 04:20:50.424281 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:20:50 crc kubenswrapper[4806]: E1204 04:20:50.425406 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:20:58 crc kubenswrapper[4806]: I1204 04:20:58.763082 4806 scope.go:117] "RemoveContainer" containerID="96112656a08459e761509553df8027474648735003ea0089e07404ec72b6f05c" Dec 04 04:20:58 crc kubenswrapper[4806]: I1204 04:20:58.786778 4806 scope.go:117] "RemoveContainer" containerID="f8ec359ef9aff652843fea70fa0ad73ef5db4188b5bc4cb5834e8e3d21ba8191" Dec 04 04:21:01 crc kubenswrapper[4806]: I1204 04:21:01.423728 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:21:01 crc kubenswrapper[4806]: E1204 04:21:01.424483 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:21:12 crc kubenswrapper[4806]: I1204 04:21:12.423486 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:21:12 crc kubenswrapper[4806]: E1204 04:21:12.424404 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:21:19 crc kubenswrapper[4806]: I1204 04:21:19.841873 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p86kf"] Dec 04 04:21:19 crc kubenswrapper[4806]: I1204 04:21:19.851605 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:19 crc kubenswrapper[4806]: I1204 04:21:19.853546 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p86kf"] Dec 04 04:21:19 crc kubenswrapper[4806]: I1204 04:21:19.959989 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbbc94e4-562c-493e-9ed2-0bd380e00249-catalog-content\") pod \"community-operators-p86kf\" (UID: \"cbbc94e4-562c-493e-9ed2-0bd380e00249\") " pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:19 crc kubenswrapper[4806]: I1204 04:21:19.960403 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbbc94e4-562c-493e-9ed2-0bd380e00249-utilities\") pod \"community-operators-p86kf\" (UID: \"cbbc94e4-562c-493e-9ed2-0bd380e00249\") " pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:19 crc kubenswrapper[4806]: I1204 04:21:19.960655 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr964\" (UniqueName: \"kubernetes.io/projected/cbbc94e4-562c-493e-9ed2-0bd380e00249-kube-api-access-jr964\") pod \"community-operators-p86kf\" (UID: \"cbbc94e4-562c-493e-9ed2-0bd380e00249\") " pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:20 crc kubenswrapper[4806]: I1204 04:21:20.062189 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr964\" (UniqueName: \"kubernetes.io/projected/cbbc94e4-562c-493e-9ed2-0bd380e00249-kube-api-access-jr964\") pod \"community-operators-p86kf\" (UID: \"cbbc94e4-562c-493e-9ed2-0bd380e00249\") " pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:20 crc kubenswrapper[4806]: I1204 04:21:20.062359 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbbc94e4-562c-493e-9ed2-0bd380e00249-catalog-content\") pod \"community-operators-p86kf\" (UID: \"cbbc94e4-562c-493e-9ed2-0bd380e00249\") " pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:20 crc kubenswrapper[4806]: I1204 04:21:20.062418 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbbc94e4-562c-493e-9ed2-0bd380e00249-utilities\") pod \"community-operators-p86kf\" (UID: \"cbbc94e4-562c-493e-9ed2-0bd380e00249\") " pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:20 crc kubenswrapper[4806]: I1204 04:21:20.063117 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbbc94e4-562c-493e-9ed2-0bd380e00249-catalog-content\") pod \"community-operators-p86kf\" (UID: \"cbbc94e4-562c-493e-9ed2-0bd380e00249\") " pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:20 crc kubenswrapper[4806]: I1204 04:21:20.063172 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbbc94e4-562c-493e-9ed2-0bd380e00249-utilities\") pod \"community-operators-p86kf\" (UID: \"cbbc94e4-562c-493e-9ed2-0bd380e00249\") " pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:20 crc kubenswrapper[4806]: I1204 04:21:20.090155 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr964\" (UniqueName: \"kubernetes.io/projected/cbbc94e4-562c-493e-9ed2-0bd380e00249-kube-api-access-jr964\") pod \"community-operators-p86kf\" (UID: \"cbbc94e4-562c-493e-9ed2-0bd380e00249\") " pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:20 crc kubenswrapper[4806]: I1204 04:21:20.173183 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:20 crc kubenswrapper[4806]: I1204 04:21:20.827596 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p86kf"] Dec 04 04:21:20 crc kubenswrapper[4806]: I1204 04:21:20.906121 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p86kf" event={"ID":"cbbc94e4-562c-493e-9ed2-0bd380e00249","Type":"ContainerStarted","Data":"debd63798e25de5da6dddc5ba908ea1dc5d5a88f45788538f9edf3fe807e52d7"} Dec 04 04:21:21 crc kubenswrapper[4806]: I1204 04:21:21.915095 4806 generic.go:334] "Generic (PLEG): container finished" podID="cbbc94e4-562c-493e-9ed2-0bd380e00249" containerID="1848d5c8750ca39b5039259722bd69d9db4089c63692c7c54c9d6de2cfe501c0" exitCode=0 Dec 04 04:21:21 crc kubenswrapper[4806]: I1204 04:21:21.915193 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p86kf" event={"ID":"cbbc94e4-562c-493e-9ed2-0bd380e00249","Type":"ContainerDied","Data":"1848d5c8750ca39b5039259722bd69d9db4089c63692c7c54c9d6de2cfe501c0"} Dec 04 04:21:22 crc kubenswrapper[4806]: I1204 04:21:22.928417 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p86kf" event={"ID":"cbbc94e4-562c-493e-9ed2-0bd380e00249","Type":"ContainerStarted","Data":"334242d08f516b530243508fb01458ed488d19b3fbb9fa37a74b23ee9faf681d"} Dec 04 04:21:24 crc kubenswrapper[4806]: I1204 04:21:24.959704 4806 generic.go:334] "Generic (PLEG): container finished" podID="cbbc94e4-562c-493e-9ed2-0bd380e00249" containerID="334242d08f516b530243508fb01458ed488d19b3fbb9fa37a74b23ee9faf681d" exitCode=0 Dec 04 04:21:24 crc kubenswrapper[4806]: I1204 04:21:24.962238 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p86kf" event={"ID":"cbbc94e4-562c-493e-9ed2-0bd380e00249","Type":"ContainerDied","Data":"334242d08f516b530243508fb01458ed488d19b3fbb9fa37a74b23ee9faf681d"} Dec 04 04:21:25 crc kubenswrapper[4806]: I1204 04:21:25.438043 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:21:25 crc kubenswrapper[4806]: E1204 04:21:25.438471 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:21:25 crc kubenswrapper[4806]: I1204 04:21:25.445952 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hktj9"] Dec 04 04:21:25 crc kubenswrapper[4806]: I1204 04:21:25.448685 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:25 crc kubenswrapper[4806]: I1204 04:21:25.455073 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hktj9"] Dec 04 04:21:25 crc kubenswrapper[4806]: I1204 04:21:25.483793 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-utilities\") pod \"redhat-marketplace-hktj9\" (UID: \"e72a2584-2822-4fba-a6dc-e6a9f29b82a3\") " pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:25 crc kubenswrapper[4806]: I1204 04:21:25.483896 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7hm9\" (UniqueName: \"kubernetes.io/projected/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-kube-api-access-w7hm9\") pod \"redhat-marketplace-hktj9\" (UID: \"e72a2584-2822-4fba-a6dc-e6a9f29b82a3\") " pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:25 crc kubenswrapper[4806]: I1204 04:21:25.484062 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-catalog-content\") pod \"redhat-marketplace-hktj9\" (UID: \"e72a2584-2822-4fba-a6dc-e6a9f29b82a3\") " pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:25 crc kubenswrapper[4806]: I1204 04:21:25.585796 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7hm9\" (UniqueName: \"kubernetes.io/projected/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-kube-api-access-w7hm9\") pod \"redhat-marketplace-hktj9\" (UID: \"e72a2584-2822-4fba-a6dc-e6a9f29b82a3\") " pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:25 crc kubenswrapper[4806]: I1204 04:21:25.586177 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-catalog-content\") pod \"redhat-marketplace-hktj9\" (UID: \"e72a2584-2822-4fba-a6dc-e6a9f29b82a3\") " pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:25 crc kubenswrapper[4806]: I1204 04:21:25.586339 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-utilities\") pod \"redhat-marketplace-hktj9\" (UID: \"e72a2584-2822-4fba-a6dc-e6a9f29b82a3\") " pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:25 crc kubenswrapper[4806]: I1204 04:21:25.587132 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-catalog-content\") pod \"redhat-marketplace-hktj9\" (UID: \"e72a2584-2822-4fba-a6dc-e6a9f29b82a3\") " pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:25 crc kubenswrapper[4806]: I1204 04:21:25.587358 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-utilities\") pod \"redhat-marketplace-hktj9\" (UID: \"e72a2584-2822-4fba-a6dc-e6a9f29b82a3\") " pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:25 crc kubenswrapper[4806]: I1204 04:21:25.612459 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7hm9\" (UniqueName: \"kubernetes.io/projected/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-kube-api-access-w7hm9\") pod \"redhat-marketplace-hktj9\" (UID: \"e72a2584-2822-4fba-a6dc-e6a9f29b82a3\") " pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:25 crc kubenswrapper[4806]: I1204 04:21:25.784078 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:26 crc kubenswrapper[4806]: I1204 04:21:26.017979 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p86kf" event={"ID":"cbbc94e4-562c-493e-9ed2-0bd380e00249","Type":"ContainerStarted","Data":"9bc33f5faee8707db166b372a0d2f70ccf3c9b24f45383497601c007cff3a38c"} Dec 04 04:21:26 crc kubenswrapper[4806]: I1204 04:21:26.395404 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p86kf" podStartSLOduration=3.962764194 podStartE2EDuration="7.395384065s" podCreationTimestamp="2025-12-04 04:21:19 +0000 UTC" firstStartedPulling="2025-12-04 04:21:21.917735376 +0000 UTC m=+1596.776248304" lastFinishedPulling="2025-12-04 04:21:25.350355247 +0000 UTC m=+1600.208868175" observedRunningTime="2025-12-04 04:21:26.054429323 +0000 UTC m=+1600.912942261" watchObservedRunningTime="2025-12-04 04:21:26.395384065 +0000 UTC m=+1601.253896993" Dec 04 04:21:26 crc kubenswrapper[4806]: I1204 04:21:26.420229 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hktj9"] Dec 04 04:21:27 crc kubenswrapper[4806]: I1204 04:21:27.027834 4806 generic.go:334] "Generic (PLEG): container finished" podID="e72a2584-2822-4fba-a6dc-e6a9f29b82a3" containerID="8bc71659801cedaacf4f3c954b5f92dcdf3e2586dd0a91b55b7c3c84d9a83eac" exitCode=0 Dec 04 04:21:27 crc kubenswrapper[4806]: I1204 04:21:27.027887 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hktj9" event={"ID":"e72a2584-2822-4fba-a6dc-e6a9f29b82a3","Type":"ContainerDied","Data":"8bc71659801cedaacf4f3c954b5f92dcdf3e2586dd0a91b55b7c3c84d9a83eac"} Dec 04 04:21:27 crc kubenswrapper[4806]: I1204 04:21:27.027953 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hktj9" event={"ID":"e72a2584-2822-4fba-a6dc-e6a9f29b82a3","Type":"ContainerStarted","Data":"6340d9cae0fb7e5a31744df13d239462730b20602daf5dde1c94e69733ab11aa"} Dec 04 04:21:27 crc kubenswrapper[4806]: I1204 04:21:27.031462 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 04:21:28 crc kubenswrapper[4806]: I1204 04:21:28.039217 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hktj9" event={"ID":"e72a2584-2822-4fba-a6dc-e6a9f29b82a3","Type":"ContainerStarted","Data":"90e3c430ff661eff83885f23dc39452df014dbceb6c63769307a611786a84192"} Dec 04 04:21:29 crc kubenswrapper[4806]: I1204 04:21:29.052303 4806 generic.go:334] "Generic (PLEG): container finished" podID="e72a2584-2822-4fba-a6dc-e6a9f29b82a3" containerID="90e3c430ff661eff83885f23dc39452df014dbceb6c63769307a611786a84192" exitCode=0 Dec 04 04:21:29 crc kubenswrapper[4806]: I1204 04:21:29.052676 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hktj9" event={"ID":"e72a2584-2822-4fba-a6dc-e6a9f29b82a3","Type":"ContainerDied","Data":"90e3c430ff661eff83885f23dc39452df014dbceb6c63769307a611786a84192"} Dec 04 04:21:30 crc kubenswrapper[4806]: I1204 04:21:30.176153 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:30 crc kubenswrapper[4806]: I1204 04:21:30.176658 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:30 crc kubenswrapper[4806]: I1204 04:21:30.248308 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:31 crc kubenswrapper[4806]: I1204 04:21:31.072825 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hktj9" event={"ID":"e72a2584-2822-4fba-a6dc-e6a9f29b82a3","Type":"ContainerStarted","Data":"da43d5a6f1ad02c3698a56ce96d50295b6b57416197b7eb3a2d8abbc22c02bd8"} Dec 04 04:21:31 crc kubenswrapper[4806]: I1204 04:21:31.108777 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hktj9" podStartSLOduration=3.102748669 podStartE2EDuration="6.108751168s" podCreationTimestamp="2025-12-04 04:21:25 +0000 UTC" firstStartedPulling="2025-12-04 04:21:27.031167067 +0000 UTC m=+1601.889679995" lastFinishedPulling="2025-12-04 04:21:30.037169566 +0000 UTC m=+1604.895682494" observedRunningTime="2025-12-04 04:21:31.09357284 +0000 UTC m=+1605.952085788" watchObservedRunningTime="2025-12-04 04:21:31.108751168 +0000 UTC m=+1605.967264106" Dec 04 04:21:31 crc kubenswrapper[4806]: I1204 04:21:31.156721 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:31 crc kubenswrapper[4806]: I1204 04:21:31.605022 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p86kf"] Dec 04 04:21:33 crc kubenswrapper[4806]: I1204 04:21:33.093839 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p86kf" podUID="cbbc94e4-562c-493e-9ed2-0bd380e00249" containerName="registry-server" containerID="cri-o://9bc33f5faee8707db166b372a0d2f70ccf3c9b24f45383497601c007cff3a38c" gracePeriod=2 Dec 04 04:21:33 crc kubenswrapper[4806]: I1204 04:21:33.522353 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:33 crc kubenswrapper[4806]: I1204 04:21:33.637275 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbbc94e4-562c-493e-9ed2-0bd380e00249-catalog-content\") pod \"cbbc94e4-562c-493e-9ed2-0bd380e00249\" (UID: \"cbbc94e4-562c-493e-9ed2-0bd380e00249\") " Dec 04 04:21:33 crc kubenswrapper[4806]: I1204 04:21:33.637633 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jr964\" (UniqueName: \"kubernetes.io/projected/cbbc94e4-562c-493e-9ed2-0bd380e00249-kube-api-access-jr964\") pod \"cbbc94e4-562c-493e-9ed2-0bd380e00249\" (UID: \"cbbc94e4-562c-493e-9ed2-0bd380e00249\") " Dec 04 04:21:33 crc kubenswrapper[4806]: I1204 04:21:33.637766 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbbc94e4-562c-493e-9ed2-0bd380e00249-utilities\") pod \"cbbc94e4-562c-493e-9ed2-0bd380e00249\" (UID: \"cbbc94e4-562c-493e-9ed2-0bd380e00249\") " Dec 04 04:21:33 crc kubenswrapper[4806]: I1204 04:21:33.638236 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbbc94e4-562c-493e-9ed2-0bd380e00249-utilities" (OuterVolumeSpecName: "utilities") pod "cbbc94e4-562c-493e-9ed2-0bd380e00249" (UID: "cbbc94e4-562c-493e-9ed2-0bd380e00249"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:21:33 crc kubenswrapper[4806]: I1204 04:21:33.638452 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cbbc94e4-562c-493e-9ed2-0bd380e00249-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:21:33 crc kubenswrapper[4806]: I1204 04:21:33.645163 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbbc94e4-562c-493e-9ed2-0bd380e00249-kube-api-access-jr964" (OuterVolumeSpecName: "kube-api-access-jr964") pod "cbbc94e4-562c-493e-9ed2-0bd380e00249" (UID: "cbbc94e4-562c-493e-9ed2-0bd380e00249"). InnerVolumeSpecName "kube-api-access-jr964". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:21:33 crc kubenswrapper[4806]: I1204 04:21:33.688563 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbbc94e4-562c-493e-9ed2-0bd380e00249-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cbbc94e4-562c-493e-9ed2-0bd380e00249" (UID: "cbbc94e4-562c-493e-9ed2-0bd380e00249"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:21:33 crc kubenswrapper[4806]: I1204 04:21:33.740688 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cbbc94e4-562c-493e-9ed2-0bd380e00249-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:21:33 crc kubenswrapper[4806]: I1204 04:21:33.740741 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jr964\" (UniqueName: \"kubernetes.io/projected/cbbc94e4-562c-493e-9ed2-0bd380e00249-kube-api-access-jr964\") on node \"crc\" DevicePath \"\"" Dec 04 04:21:34 crc kubenswrapper[4806]: I1204 04:21:34.108745 4806 generic.go:334] "Generic (PLEG): container finished" podID="cbbc94e4-562c-493e-9ed2-0bd380e00249" containerID="9bc33f5faee8707db166b372a0d2f70ccf3c9b24f45383497601c007cff3a38c" exitCode=0 Dec 04 04:21:34 crc kubenswrapper[4806]: I1204 04:21:34.108807 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p86kf" event={"ID":"cbbc94e4-562c-493e-9ed2-0bd380e00249","Type":"ContainerDied","Data":"9bc33f5faee8707db166b372a0d2f70ccf3c9b24f45383497601c007cff3a38c"} Dec 04 04:21:34 crc kubenswrapper[4806]: I1204 04:21:34.108860 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p86kf" Dec 04 04:21:34 crc kubenswrapper[4806]: I1204 04:21:34.109178 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p86kf" event={"ID":"cbbc94e4-562c-493e-9ed2-0bd380e00249","Type":"ContainerDied","Data":"debd63798e25de5da6dddc5ba908ea1dc5d5a88f45788538f9edf3fe807e52d7"} Dec 04 04:21:34 crc kubenswrapper[4806]: I1204 04:21:34.109206 4806 scope.go:117] "RemoveContainer" containerID="9bc33f5faee8707db166b372a0d2f70ccf3c9b24f45383497601c007cff3a38c" Dec 04 04:21:34 crc kubenswrapper[4806]: I1204 04:21:34.160691 4806 scope.go:117] "RemoveContainer" containerID="334242d08f516b530243508fb01458ed488d19b3fbb9fa37a74b23ee9faf681d" Dec 04 04:21:34 crc kubenswrapper[4806]: I1204 04:21:34.162208 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p86kf"] Dec 04 04:21:34 crc kubenswrapper[4806]: I1204 04:21:34.173604 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p86kf"] Dec 04 04:21:34 crc kubenswrapper[4806]: I1204 04:21:34.224981 4806 scope.go:117] "RemoveContainer" containerID="1848d5c8750ca39b5039259722bd69d9db4089c63692c7c54c9d6de2cfe501c0" Dec 04 04:21:34 crc kubenswrapper[4806]: I1204 04:21:34.266294 4806 scope.go:117] "RemoveContainer" containerID="9bc33f5faee8707db166b372a0d2f70ccf3c9b24f45383497601c007cff3a38c" Dec 04 04:21:34 crc kubenswrapper[4806]: E1204 04:21:34.266769 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bc33f5faee8707db166b372a0d2f70ccf3c9b24f45383497601c007cff3a38c\": container with ID starting with 9bc33f5faee8707db166b372a0d2f70ccf3c9b24f45383497601c007cff3a38c not found: ID does not exist" containerID="9bc33f5faee8707db166b372a0d2f70ccf3c9b24f45383497601c007cff3a38c" Dec 04 04:21:34 crc kubenswrapper[4806]: I1204 04:21:34.266801 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bc33f5faee8707db166b372a0d2f70ccf3c9b24f45383497601c007cff3a38c"} err="failed to get container status \"9bc33f5faee8707db166b372a0d2f70ccf3c9b24f45383497601c007cff3a38c\": rpc error: code = NotFound desc = could not find container \"9bc33f5faee8707db166b372a0d2f70ccf3c9b24f45383497601c007cff3a38c\": container with ID starting with 9bc33f5faee8707db166b372a0d2f70ccf3c9b24f45383497601c007cff3a38c not found: ID does not exist" Dec 04 04:21:34 crc kubenswrapper[4806]: I1204 04:21:34.266820 4806 scope.go:117] "RemoveContainer" containerID="334242d08f516b530243508fb01458ed488d19b3fbb9fa37a74b23ee9faf681d" Dec 04 04:21:34 crc kubenswrapper[4806]: E1204 04:21:34.267218 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"334242d08f516b530243508fb01458ed488d19b3fbb9fa37a74b23ee9faf681d\": container with ID starting with 334242d08f516b530243508fb01458ed488d19b3fbb9fa37a74b23ee9faf681d not found: ID does not exist" containerID="334242d08f516b530243508fb01458ed488d19b3fbb9fa37a74b23ee9faf681d" Dec 04 04:21:34 crc kubenswrapper[4806]: I1204 04:21:34.267246 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"334242d08f516b530243508fb01458ed488d19b3fbb9fa37a74b23ee9faf681d"} err="failed to get container status \"334242d08f516b530243508fb01458ed488d19b3fbb9fa37a74b23ee9faf681d\": rpc error: code = NotFound desc = could not find container \"334242d08f516b530243508fb01458ed488d19b3fbb9fa37a74b23ee9faf681d\": container with ID starting with 334242d08f516b530243508fb01458ed488d19b3fbb9fa37a74b23ee9faf681d not found: ID does not exist" Dec 04 04:21:34 crc kubenswrapper[4806]: I1204 04:21:34.267260 4806 scope.go:117] "RemoveContainer" containerID="1848d5c8750ca39b5039259722bd69d9db4089c63692c7c54c9d6de2cfe501c0" Dec 04 04:21:34 crc kubenswrapper[4806]: E1204 04:21:34.267478 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1848d5c8750ca39b5039259722bd69d9db4089c63692c7c54c9d6de2cfe501c0\": container with ID starting with 1848d5c8750ca39b5039259722bd69d9db4089c63692c7c54c9d6de2cfe501c0 not found: ID does not exist" containerID="1848d5c8750ca39b5039259722bd69d9db4089c63692c7c54c9d6de2cfe501c0" Dec 04 04:21:34 crc kubenswrapper[4806]: I1204 04:21:34.267498 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1848d5c8750ca39b5039259722bd69d9db4089c63692c7c54c9d6de2cfe501c0"} err="failed to get container status \"1848d5c8750ca39b5039259722bd69d9db4089c63692c7c54c9d6de2cfe501c0\": rpc error: code = NotFound desc = could not find container \"1848d5c8750ca39b5039259722bd69d9db4089c63692c7c54c9d6de2cfe501c0\": container with ID starting with 1848d5c8750ca39b5039259722bd69d9db4089c63692c7c54c9d6de2cfe501c0 not found: ID does not exist" Dec 04 04:21:35 crc kubenswrapper[4806]: I1204 04:21:35.435203 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbbc94e4-562c-493e-9ed2-0bd380e00249" path="/var/lib/kubelet/pods/cbbc94e4-562c-493e-9ed2-0bd380e00249/volumes" Dec 04 04:21:35 crc kubenswrapper[4806]: I1204 04:21:35.786178 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:35 crc kubenswrapper[4806]: I1204 04:21:35.786571 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:35 crc kubenswrapper[4806]: I1204 04:21:35.834976 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:36 crc kubenswrapper[4806]: I1204 04:21:36.252237 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:36 crc kubenswrapper[4806]: I1204 04:21:36.807140 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hktj9"] Dec 04 04:21:38 crc kubenswrapper[4806]: I1204 04:21:38.144508 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hktj9" podUID="e72a2584-2822-4fba-a6dc-e6a9f29b82a3" containerName="registry-server" containerID="cri-o://da43d5a6f1ad02c3698a56ce96d50295b6b57416197b7eb3a2d8abbc22c02bd8" gracePeriod=2 Dec 04 04:21:38 crc kubenswrapper[4806]: I1204 04:21:38.576682 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:38 crc kubenswrapper[4806]: I1204 04:21:38.627443 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-catalog-content\") pod \"e72a2584-2822-4fba-a6dc-e6a9f29b82a3\" (UID: \"e72a2584-2822-4fba-a6dc-e6a9f29b82a3\") " Dec 04 04:21:38 crc kubenswrapper[4806]: I1204 04:21:38.627486 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-utilities\") pod \"e72a2584-2822-4fba-a6dc-e6a9f29b82a3\" (UID: \"e72a2584-2822-4fba-a6dc-e6a9f29b82a3\") " Dec 04 04:21:38 crc kubenswrapper[4806]: I1204 04:21:38.627544 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7hm9\" (UniqueName: \"kubernetes.io/projected/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-kube-api-access-w7hm9\") pod \"e72a2584-2822-4fba-a6dc-e6a9f29b82a3\" (UID: \"e72a2584-2822-4fba-a6dc-e6a9f29b82a3\") " Dec 04 04:21:38 crc kubenswrapper[4806]: I1204 04:21:38.629716 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-utilities" (OuterVolumeSpecName: "utilities") pod "e72a2584-2822-4fba-a6dc-e6a9f29b82a3" (UID: "e72a2584-2822-4fba-a6dc-e6a9f29b82a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:21:38 crc kubenswrapper[4806]: I1204 04:21:38.639211 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-kube-api-access-w7hm9" (OuterVolumeSpecName: "kube-api-access-w7hm9") pod "e72a2584-2822-4fba-a6dc-e6a9f29b82a3" (UID: "e72a2584-2822-4fba-a6dc-e6a9f29b82a3"). InnerVolumeSpecName "kube-api-access-w7hm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:21:38 crc kubenswrapper[4806]: I1204 04:21:38.660509 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e72a2584-2822-4fba-a6dc-e6a9f29b82a3" (UID: "e72a2584-2822-4fba-a6dc-e6a9f29b82a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:21:38 crc kubenswrapper[4806]: I1204 04:21:38.730017 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:21:38 crc kubenswrapper[4806]: I1204 04:21:38.730336 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:21:38 crc kubenswrapper[4806]: I1204 04:21:38.730349 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7hm9\" (UniqueName: \"kubernetes.io/projected/e72a2584-2822-4fba-a6dc-e6a9f29b82a3-kube-api-access-w7hm9\") on node \"crc\" DevicePath \"\"" Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.158157 4806 generic.go:334] "Generic (PLEG): container finished" podID="e72a2584-2822-4fba-a6dc-e6a9f29b82a3" containerID="da43d5a6f1ad02c3698a56ce96d50295b6b57416197b7eb3a2d8abbc22c02bd8" exitCode=0 Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.158228 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hktj9" event={"ID":"e72a2584-2822-4fba-a6dc-e6a9f29b82a3","Type":"ContainerDied","Data":"da43d5a6f1ad02c3698a56ce96d50295b6b57416197b7eb3a2d8abbc22c02bd8"} Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.158265 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hktj9" event={"ID":"e72a2584-2822-4fba-a6dc-e6a9f29b82a3","Type":"ContainerDied","Data":"6340d9cae0fb7e5a31744df13d239462730b20602daf5dde1c94e69733ab11aa"} Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.158289 4806 scope.go:117] "RemoveContainer" containerID="da43d5a6f1ad02c3698a56ce96d50295b6b57416197b7eb3a2d8abbc22c02bd8" Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.158357 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hktj9" Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.194080 4806 scope.go:117] "RemoveContainer" containerID="90e3c430ff661eff83885f23dc39452df014dbceb6c63769307a611786a84192" Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.202129 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hktj9"] Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.218418 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hktj9"] Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.240219 4806 scope.go:117] "RemoveContainer" containerID="8bc71659801cedaacf4f3c954b5f92dcdf3e2586dd0a91b55b7c3c84d9a83eac" Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.285266 4806 scope.go:117] "RemoveContainer" containerID="da43d5a6f1ad02c3698a56ce96d50295b6b57416197b7eb3a2d8abbc22c02bd8" Dec 04 04:21:39 crc kubenswrapper[4806]: E1204 04:21:39.286008 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da43d5a6f1ad02c3698a56ce96d50295b6b57416197b7eb3a2d8abbc22c02bd8\": container with ID starting with da43d5a6f1ad02c3698a56ce96d50295b6b57416197b7eb3a2d8abbc22c02bd8 not found: ID does not exist" containerID="da43d5a6f1ad02c3698a56ce96d50295b6b57416197b7eb3a2d8abbc22c02bd8" Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.286075 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da43d5a6f1ad02c3698a56ce96d50295b6b57416197b7eb3a2d8abbc22c02bd8"} err="failed to get container status \"da43d5a6f1ad02c3698a56ce96d50295b6b57416197b7eb3a2d8abbc22c02bd8\": rpc error: code = NotFound desc = could not find container \"da43d5a6f1ad02c3698a56ce96d50295b6b57416197b7eb3a2d8abbc22c02bd8\": container with ID starting with da43d5a6f1ad02c3698a56ce96d50295b6b57416197b7eb3a2d8abbc22c02bd8 not found: ID does not exist" Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.286097 4806 scope.go:117] "RemoveContainer" containerID="90e3c430ff661eff83885f23dc39452df014dbceb6c63769307a611786a84192" Dec 04 04:21:39 crc kubenswrapper[4806]: E1204 04:21:39.286867 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90e3c430ff661eff83885f23dc39452df014dbceb6c63769307a611786a84192\": container with ID starting with 90e3c430ff661eff83885f23dc39452df014dbceb6c63769307a611786a84192 not found: ID does not exist" containerID="90e3c430ff661eff83885f23dc39452df014dbceb6c63769307a611786a84192" Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.286894 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90e3c430ff661eff83885f23dc39452df014dbceb6c63769307a611786a84192"} err="failed to get container status \"90e3c430ff661eff83885f23dc39452df014dbceb6c63769307a611786a84192\": rpc error: code = NotFound desc = could not find container \"90e3c430ff661eff83885f23dc39452df014dbceb6c63769307a611786a84192\": container with ID starting with 90e3c430ff661eff83885f23dc39452df014dbceb6c63769307a611786a84192 not found: ID does not exist" Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.286914 4806 scope.go:117] "RemoveContainer" containerID="8bc71659801cedaacf4f3c954b5f92dcdf3e2586dd0a91b55b7c3c84d9a83eac" Dec 04 04:21:39 crc kubenswrapper[4806]: E1204 04:21:39.287312 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bc71659801cedaacf4f3c954b5f92dcdf3e2586dd0a91b55b7c3c84d9a83eac\": container with ID starting with 8bc71659801cedaacf4f3c954b5f92dcdf3e2586dd0a91b55b7c3c84d9a83eac not found: ID does not exist" containerID="8bc71659801cedaacf4f3c954b5f92dcdf3e2586dd0a91b55b7c3c84d9a83eac" Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.287364 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bc71659801cedaacf4f3c954b5f92dcdf3e2586dd0a91b55b7c3c84d9a83eac"} err="failed to get container status \"8bc71659801cedaacf4f3c954b5f92dcdf3e2586dd0a91b55b7c3c84d9a83eac\": rpc error: code = NotFound desc = could not find container \"8bc71659801cedaacf4f3c954b5f92dcdf3e2586dd0a91b55b7c3c84d9a83eac\": container with ID starting with 8bc71659801cedaacf4f3c954b5f92dcdf3e2586dd0a91b55b7c3c84d9a83eac not found: ID does not exist" Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.425143 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:21:39 crc kubenswrapper[4806]: E1204 04:21:39.425397 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:21:39 crc kubenswrapper[4806]: I1204 04:21:39.436620 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e72a2584-2822-4fba-a6dc-e6a9f29b82a3" path="/var/lib/kubelet/pods/e72a2584-2822-4fba-a6dc-e6a9f29b82a3/volumes" Dec 04 04:21:54 crc kubenswrapper[4806]: I1204 04:21:54.423630 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:21:54 crc kubenswrapper[4806]: E1204 04:21:54.425227 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:21:58 crc kubenswrapper[4806]: I1204 04:21:58.879360 4806 scope.go:117] "RemoveContainer" containerID="39f515cae2496194a873198e9bcc3ae7bc4668ff452c27135e6987ebf4276afc" Dec 04 04:21:58 crc kubenswrapper[4806]: I1204 04:21:58.903431 4806 scope.go:117] "RemoveContainer" containerID="8428b43f132768c89eff84a87a2ffb8f9182b7bbc7b553fed98c07bff6da1ba9" Dec 04 04:21:58 crc kubenswrapper[4806]: I1204 04:21:58.946095 4806 scope.go:117] "RemoveContainer" containerID="44c099c1b1e2667cc7a3c3a248b3234946f9dc0e882330b14185ec61de016538" Dec 04 04:21:58 crc kubenswrapper[4806]: I1204 04:21:58.967407 4806 scope.go:117] "RemoveContainer" containerID="b84ee829c765879ba892935281145d03c4ca17da2bcde8164c8b78105e71d72e" Dec 04 04:21:58 crc kubenswrapper[4806]: I1204 04:21:58.988220 4806 scope.go:117] "RemoveContainer" containerID="95dd0182bcbefacae5d7c5b8075c03f37fcb2e14fb34b4ee557b3091755a4825" Dec 04 04:22:05 crc kubenswrapper[4806]: I1204 04:22:05.430542 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:22:05 crc kubenswrapper[4806]: E1204 04:22:05.431439 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:22:20 crc kubenswrapper[4806]: I1204 04:22:20.423906 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:22:20 crc kubenswrapper[4806]: E1204 04:22:20.424792 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:22:32 crc kubenswrapper[4806]: I1204 04:22:32.423587 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:22:32 crc kubenswrapper[4806]: E1204 04:22:32.424567 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:22:46 crc kubenswrapper[4806]: I1204 04:22:46.425192 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:22:46 crc kubenswrapper[4806]: E1204 04:22:46.426458 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:23:01 crc kubenswrapper[4806]: I1204 04:23:01.424366 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:23:01 crc kubenswrapper[4806]: E1204 04:23:01.425172 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:23:14 crc kubenswrapper[4806]: I1204 04:23:14.424978 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:23:14 crc kubenswrapper[4806]: E1204 04:23:14.425891 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:23:16 crc kubenswrapper[4806]: I1204 04:23:16.041620 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-bd6f-account-create-update-xk87b"] Dec 04 04:23:16 crc kubenswrapper[4806]: I1204 04:23:16.052105 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-bd6f-account-create-update-xk87b"] Dec 04 04:23:17 crc kubenswrapper[4806]: I1204 04:23:17.047862 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-mz5hn"] Dec 04 04:23:17 crc kubenswrapper[4806]: I1204 04:23:17.059221 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-mz5hn"] Dec 04 04:23:17 crc kubenswrapper[4806]: I1204 04:23:17.434454 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="361a7154-0c20-4c88-bedc-f2bb0ea4d08d" path="/var/lib/kubelet/pods/361a7154-0c20-4c88-bedc-f2bb0ea4d08d/volumes" Dec 04 04:23:17 crc kubenswrapper[4806]: I1204 04:23:17.436579 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3da131f7-72d9-4cc8-bc7b-f361b456f08c" path="/var/lib/kubelet/pods/3da131f7-72d9-4cc8-bc7b-f361b456f08c/volumes" Dec 04 04:23:25 crc kubenswrapper[4806]: I1204 04:23:25.041729 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-tjvgt"] Dec 04 04:23:25 crc kubenswrapper[4806]: I1204 04:23:25.059056 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-122a-account-create-update-nxhgt"] Dec 04 04:23:25 crc kubenswrapper[4806]: I1204 04:23:25.072155 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-122a-account-create-update-nxhgt"] Dec 04 04:23:25 crc kubenswrapper[4806]: I1204 04:23:25.081782 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-tjvgt"] Dec 04 04:23:25 crc kubenswrapper[4806]: I1204 04:23:25.434314 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc7a00c7-53ea-4717-89b1-1266f50590d0" path="/var/lib/kubelet/pods/cc7a00c7-53ea-4717-89b1-1266f50590d0/volumes" Dec 04 04:23:25 crc kubenswrapper[4806]: I1204 04:23:25.435058 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cea7a0d7-7a20-4c9a-90fd-8bd779be02b9" path="/var/lib/kubelet/pods/cea7a0d7-7a20-4c9a-90fd-8bd779be02b9/volumes" Dec 04 04:23:26 crc kubenswrapper[4806]: I1204 04:23:26.424132 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:23:26 crc kubenswrapper[4806]: E1204 04:23:26.424388 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:23:28 crc kubenswrapper[4806]: I1204 04:23:28.038202 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-wjzmn"] Dec 04 04:23:28 crc kubenswrapper[4806]: I1204 04:23:28.048946 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-9fa4-account-create-update-2pkcm"] Dec 04 04:23:28 crc kubenswrapper[4806]: I1204 04:23:28.060236 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-9fa4-account-create-update-2pkcm"] Dec 04 04:23:28 crc kubenswrapper[4806]: I1204 04:23:28.070185 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-wjzmn"] Dec 04 04:23:29 crc kubenswrapper[4806]: I1204 04:23:29.436339 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a872a38c-7d3d-4685-aed2-b224c9f18a4a" path="/var/lib/kubelet/pods/a872a38c-7d3d-4685-aed2-b224c9f18a4a/volumes" Dec 04 04:23:29 crc kubenswrapper[4806]: I1204 04:23:29.438682 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7d42f80-97fb-4174-b256-ccf414215b24" path="/var/lib/kubelet/pods/c7d42f80-97fb-4174-b256-ccf414215b24/volumes" Dec 04 04:23:40 crc kubenswrapper[4806]: I1204 04:23:40.424604 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:23:40 crc kubenswrapper[4806]: E1204 04:23:40.426262 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:23:54 crc kubenswrapper[4806]: I1204 04:23:54.424175 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:23:54 crc kubenswrapper[4806]: E1204 04:23:54.425114 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:23:55 crc kubenswrapper[4806]: I1204 04:23:55.047420 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-pgg86"] Dec 04 04:23:55 crc kubenswrapper[4806]: I1204 04:23:55.056838 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-pgg86"] Dec 04 04:23:55 crc kubenswrapper[4806]: I1204 04:23:55.435143 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d576c39-2a71-4977-afab-4d1746f7c38f" path="/var/lib/kubelet/pods/3d576c39-2a71-4977-afab-4d1746f7c38f/volumes" Dec 04 04:23:56 crc kubenswrapper[4806]: I1204 04:23:56.040544 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-2gm7t"] Dec 04 04:23:56 crc kubenswrapper[4806]: I1204 04:23:56.075262 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-0d3f-account-create-update-lkzhq"] Dec 04 04:23:56 crc kubenswrapper[4806]: I1204 04:23:56.084882 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-2gm7t"] Dec 04 04:23:56 crc kubenswrapper[4806]: I1204 04:23:56.094143 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-s5fdn"] Dec 04 04:23:56 crc kubenswrapper[4806]: I1204 04:23:56.103163 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-s5fdn"] Dec 04 04:23:56 crc kubenswrapper[4806]: I1204 04:23:56.112057 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-0d3f-account-create-update-lkzhq"] Dec 04 04:23:57 crc kubenswrapper[4806]: I1204 04:23:57.030065 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6ba4-account-create-update-nzmnw"] Dec 04 04:23:57 crc kubenswrapper[4806]: I1204 04:23:57.039714 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-e2df-account-create-update-ch7v6"] Dec 04 04:23:57 crc kubenswrapper[4806]: I1204 04:23:57.049937 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6ba4-account-create-update-nzmnw"] Dec 04 04:23:57 crc kubenswrapper[4806]: I1204 04:23:57.059789 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-e2df-account-create-update-ch7v6"] Dec 04 04:23:57 crc kubenswrapper[4806]: I1204 04:23:57.438211 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7199166b-2234-4b45-b086-3129a62fc2fc" path="/var/lib/kubelet/pods/7199166b-2234-4b45-b086-3129a62fc2fc/volumes" Dec 04 04:23:57 crc kubenswrapper[4806]: I1204 04:23:57.440198 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a41b88c9-7bfc-47e8-b612-7129c0821be3" path="/var/lib/kubelet/pods/a41b88c9-7bfc-47e8-b612-7129c0821be3/volumes" Dec 04 04:23:57 crc kubenswrapper[4806]: I1204 04:23:57.441915 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be066c54-140f-4727-88eb-864a9d366de5" path="/var/lib/kubelet/pods/be066c54-140f-4727-88eb-864a9d366de5/volumes" Dec 04 04:23:57 crc kubenswrapper[4806]: I1204 04:23:57.443018 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea145431-aaa9-4fee-8485-bf510144f89e" path="/var/lib/kubelet/pods/ea145431-aaa9-4fee-8485-bf510144f89e/volumes" Dec 04 04:23:57 crc kubenswrapper[4806]: I1204 04:23:57.444501 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb787acc-3d9f-43ea-a016-0232e3bd1750" path="/var/lib/kubelet/pods/fb787acc-3d9f-43ea-a016-0232e3bd1750/volumes" Dec 04 04:23:59 crc kubenswrapper[4806]: I1204 04:23:59.192968 4806 scope.go:117] "RemoveContainer" containerID="dc3178cb5098590a65febd9a7c27176e1b7a85beeea2f546c10697bffa660e52" Dec 04 04:23:59 crc kubenswrapper[4806]: I1204 04:23:59.322166 4806 scope.go:117] "RemoveContainer" containerID="0d00e7796d93a7ceb9886b3707fee3a8baa16c9e4d7883cf1312744e209b701e" Dec 04 04:23:59 crc kubenswrapper[4806]: I1204 04:23:59.396093 4806 scope.go:117] "RemoveContainer" containerID="004d8da71d4f9d7ffb858dde4d8f7e94b4be954cbcb1a1d9b5a7453542420111" Dec 04 04:23:59 crc kubenswrapper[4806]: I1204 04:23:59.465147 4806 scope.go:117] "RemoveContainer" containerID="33e9f1443117789b81d7d1a6d67bc35ac21a29cf81a9a9af687cbfd2b17e9eaf" Dec 04 04:23:59 crc kubenswrapper[4806]: I1204 04:23:59.532621 4806 scope.go:117] "RemoveContainer" containerID="95ed62f0bece6e468c6fbabe37b8c61afc08b48b8c6176bb714a60fd11274810" Dec 04 04:23:59 crc kubenswrapper[4806]: I1204 04:23:59.605327 4806 scope.go:117] "RemoveContainer" containerID="0bb85df53d42f9e7a74f60699fb3a94b3aa5b76911eb3437eb2ea9c79eaf6765" Dec 04 04:23:59 crc kubenswrapper[4806]: I1204 04:23:59.677182 4806 scope.go:117] "RemoveContainer" containerID="8c63f8abe1919c41f090e80234924f24282b370031e3e0726293ff3c3921a983" Dec 04 04:23:59 crc kubenswrapper[4806]: I1204 04:23:59.712601 4806 scope.go:117] "RemoveContainer" containerID="be588154a943039e689938cd9ed3e06c2255fef9793cb887b6feb992cd644828" Dec 04 04:23:59 crc kubenswrapper[4806]: I1204 04:23:59.760471 4806 scope.go:117] "RemoveContainer" containerID="e205d21ce56305706301333692cf1d160005ed077a450f8885eafc29f5df3ded" Dec 04 04:23:59 crc kubenswrapper[4806]: I1204 04:23:59.781680 4806 scope.go:117] "RemoveContainer" containerID="4dd482f163c77b04bf138f255793d02c3127369b5cb07280687282da52d0e5e0" Dec 04 04:23:59 crc kubenswrapper[4806]: I1204 04:23:59.803597 4806 scope.go:117] "RemoveContainer" containerID="d4c0f4f92bc396d80d312491155fad120886c6711e7f693b14767625c170833c" Dec 04 04:23:59 crc kubenswrapper[4806]: I1204 04:23:59.839526 4806 scope.go:117] "RemoveContainer" containerID="40b75f909ea885f4daa09f616dc25bdd594fe94ddb0d92f924e73b20320cd3ac" Dec 04 04:23:59 crc kubenswrapper[4806]: I1204 04:23:59.873568 4806 scope.go:117] "RemoveContainer" containerID="df2248cd6cef99d798274e361bbb050ff8710003fca0f80a0e1ee6e16aa4ae2b" Dec 04 04:23:59 crc kubenswrapper[4806]: I1204 04:23:59.901180 4806 scope.go:117] "RemoveContainer" containerID="4d8b7a947762eae6934fd38b651c52659cb7dec30803846fa7e3a5b167146b93" Dec 04 04:24:05 crc kubenswrapper[4806]: I1204 04:24:05.433272 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:24:05 crc kubenswrapper[4806]: E1204 04:24:05.434201 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:24:09 crc kubenswrapper[4806]: I1204 04:24:09.029863 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-hbvv2"] Dec 04 04:24:09 crc kubenswrapper[4806]: I1204 04:24:09.040061 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-hbvv2"] Dec 04 04:24:09 crc kubenswrapper[4806]: I1204 04:24:09.439747 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b10867b-8469-4116-ba55-f6a7cc7b8e01" path="/var/lib/kubelet/pods/3b10867b-8469-4116-ba55-f6a7cc7b8e01/volumes" Dec 04 04:24:10 crc kubenswrapper[4806]: I1204 04:24:10.519962 4806 generic.go:334] "Generic (PLEG): container finished" podID="1787b37a-3067-4912-a658-4a8fd53faab6" containerID="bf9ed1e2f3a08af0018cf98babc8fa1772a06f8599d234b8a0f4c6bc97bc3e32" exitCode=0 Dec 04 04:24:10 crc kubenswrapper[4806]: I1204 04:24:10.520177 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" event={"ID":"1787b37a-3067-4912-a658-4a8fd53faab6","Type":"ContainerDied","Data":"bf9ed1e2f3a08af0018cf98babc8fa1772a06f8599d234b8a0f4c6bc97bc3e32"} Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.001701 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.147201 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-bootstrap-combined-ca-bundle\") pod \"1787b37a-3067-4912-a658-4a8fd53faab6\" (UID: \"1787b37a-3067-4912-a658-4a8fd53faab6\") " Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.147888 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hd4r9\" (UniqueName: \"kubernetes.io/projected/1787b37a-3067-4912-a658-4a8fd53faab6-kube-api-access-hd4r9\") pod \"1787b37a-3067-4912-a658-4a8fd53faab6\" (UID: \"1787b37a-3067-4912-a658-4a8fd53faab6\") " Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.148027 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-inventory\") pod \"1787b37a-3067-4912-a658-4a8fd53faab6\" (UID: \"1787b37a-3067-4912-a658-4a8fd53faab6\") " Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.148215 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-ssh-key\") pod \"1787b37a-3067-4912-a658-4a8fd53faab6\" (UID: \"1787b37a-3067-4912-a658-4a8fd53faab6\") " Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.159954 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1787b37a-3067-4912-a658-4a8fd53faab6-kube-api-access-hd4r9" (OuterVolumeSpecName: "kube-api-access-hd4r9") pod "1787b37a-3067-4912-a658-4a8fd53faab6" (UID: "1787b37a-3067-4912-a658-4a8fd53faab6"). InnerVolumeSpecName "kube-api-access-hd4r9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.160290 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "1787b37a-3067-4912-a658-4a8fd53faab6" (UID: "1787b37a-3067-4912-a658-4a8fd53faab6"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.191874 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1787b37a-3067-4912-a658-4a8fd53faab6" (UID: "1787b37a-3067-4912-a658-4a8fd53faab6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.194207 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-inventory" (OuterVolumeSpecName: "inventory") pod "1787b37a-3067-4912-a658-4a8fd53faab6" (UID: "1787b37a-3067-4912-a658-4a8fd53faab6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.250450 4806 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.250490 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hd4r9\" (UniqueName: \"kubernetes.io/projected/1787b37a-3067-4912-a658-4a8fd53faab6-kube-api-access-hd4r9\") on node \"crc\" DevicePath \"\"" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.250501 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.250512 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1787b37a-3067-4912-a658-4a8fd53faab6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.543762 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" event={"ID":"1787b37a-3067-4912-a658-4a8fd53faab6","Type":"ContainerDied","Data":"b2d7947e5bb91eeb123dfdbd555cc1ede902a15414fbdd68018c9df3a131c789"} Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.543831 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2d7947e5bb91eeb123dfdbd555cc1ede902a15414fbdd68018c9df3a131c789" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.543951 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.635217 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv"] Dec 04 04:24:12 crc kubenswrapper[4806]: E1204 04:24:12.635595 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbbc94e4-562c-493e-9ed2-0bd380e00249" containerName="extract-utilities" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.635621 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbbc94e4-562c-493e-9ed2-0bd380e00249" containerName="extract-utilities" Dec 04 04:24:12 crc kubenswrapper[4806]: E1204 04:24:12.635643 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1787b37a-3067-4912-a658-4a8fd53faab6" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.635650 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="1787b37a-3067-4912-a658-4a8fd53faab6" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 04 04:24:12 crc kubenswrapper[4806]: E1204 04:24:12.635664 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbbc94e4-562c-493e-9ed2-0bd380e00249" containerName="extract-content" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.635671 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbbc94e4-562c-493e-9ed2-0bd380e00249" containerName="extract-content" Dec 04 04:24:12 crc kubenswrapper[4806]: E1204 04:24:12.635683 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e72a2584-2822-4fba-a6dc-e6a9f29b82a3" containerName="extract-content" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.635689 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e72a2584-2822-4fba-a6dc-e6a9f29b82a3" containerName="extract-content" Dec 04 04:24:12 crc kubenswrapper[4806]: E1204 04:24:12.635701 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e72a2584-2822-4fba-a6dc-e6a9f29b82a3" containerName="extract-utilities" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.635707 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e72a2584-2822-4fba-a6dc-e6a9f29b82a3" containerName="extract-utilities" Dec 04 04:24:12 crc kubenswrapper[4806]: E1204 04:24:12.635722 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbbc94e4-562c-493e-9ed2-0bd380e00249" containerName="registry-server" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.635728 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbbc94e4-562c-493e-9ed2-0bd380e00249" containerName="registry-server" Dec 04 04:24:12 crc kubenswrapper[4806]: E1204 04:24:12.635744 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e72a2584-2822-4fba-a6dc-e6a9f29b82a3" containerName="registry-server" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.635751 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e72a2584-2822-4fba-a6dc-e6a9f29b82a3" containerName="registry-server" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.635942 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e72a2584-2822-4fba-a6dc-e6a9f29b82a3" containerName="registry-server" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.635974 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="1787b37a-3067-4912-a658-4a8fd53faab6" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.635995 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbbc94e4-562c-493e-9ed2-0bd380e00249" containerName="registry-server" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.636754 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.640372 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.640612 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.640780 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-q8mqk" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.641999 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.659904 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv"] Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.764984 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9ll2\" (UniqueName: \"kubernetes.io/projected/0b8e6232-8e76-425c-9b02-94e0a3c236cb-kube-api-access-l9ll2\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv\" (UID: \"0b8e6232-8e76-425c-9b02-94e0a3c236cb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.765099 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b8e6232-8e76-425c-9b02-94e0a3c236cb-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv\" (UID: \"0b8e6232-8e76-425c-9b02-94e0a3c236cb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.765132 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b8e6232-8e76-425c-9b02-94e0a3c236cb-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv\" (UID: \"0b8e6232-8e76-425c-9b02-94e0a3c236cb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.866549 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b8e6232-8e76-425c-9b02-94e0a3c236cb-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv\" (UID: \"0b8e6232-8e76-425c-9b02-94e0a3c236cb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.866797 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b8e6232-8e76-425c-9b02-94e0a3c236cb-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv\" (UID: \"0b8e6232-8e76-425c-9b02-94e0a3c236cb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.867041 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9ll2\" (UniqueName: \"kubernetes.io/projected/0b8e6232-8e76-425c-9b02-94e0a3c236cb-kube-api-access-l9ll2\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv\" (UID: \"0b8e6232-8e76-425c-9b02-94e0a3c236cb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.870681 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b8e6232-8e76-425c-9b02-94e0a3c236cb-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv\" (UID: \"0b8e6232-8e76-425c-9b02-94e0a3c236cb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.874552 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b8e6232-8e76-425c-9b02-94e0a3c236cb-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv\" (UID: \"0b8e6232-8e76-425c-9b02-94e0a3c236cb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.886387 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9ll2\" (UniqueName: \"kubernetes.io/projected/0b8e6232-8e76-425c-9b02-94e0a3c236cb-kube-api-access-l9ll2\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv\" (UID: \"0b8e6232-8e76-425c-9b02-94e0a3c236cb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" Dec 04 04:24:12 crc kubenswrapper[4806]: I1204 04:24:12.960199 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" Dec 04 04:24:13 crc kubenswrapper[4806]: I1204 04:24:13.532554 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv"] Dec 04 04:24:13 crc kubenswrapper[4806]: I1204 04:24:13.553910 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" event={"ID":"0b8e6232-8e76-425c-9b02-94e0a3c236cb","Type":"ContainerStarted","Data":"d14b593f6811ab2946ae9d0b0a3baf528fb441eb309058c8148ed57505019a43"} Dec 04 04:24:14 crc kubenswrapper[4806]: I1204 04:24:14.564985 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" event={"ID":"0b8e6232-8e76-425c-9b02-94e0a3c236cb","Type":"ContainerStarted","Data":"8514593d5c52ceae8d55eb83becaeb2a26e47f33d06cd864ab07d7f6cb1a1cde"} Dec 04 04:24:18 crc kubenswrapper[4806]: I1204 04:24:18.050600 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" podStartSLOduration=5.891949123 podStartE2EDuration="6.050570913s" podCreationTimestamp="2025-12-04 04:24:12 +0000 UTC" firstStartedPulling="2025-12-04 04:24:13.534942571 +0000 UTC m=+1768.393455499" lastFinishedPulling="2025-12-04 04:24:13.693564361 +0000 UTC m=+1768.552077289" observedRunningTime="2025-12-04 04:24:14.588614753 +0000 UTC m=+1769.447127681" watchObservedRunningTime="2025-12-04 04:24:18.050570913 +0000 UTC m=+1772.909083841" Dec 04 04:24:18 crc kubenswrapper[4806]: I1204 04:24:18.051710 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-r7mb9"] Dec 04 04:24:18 crc kubenswrapper[4806]: I1204 04:24:18.062840 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-r7mb9"] Dec 04 04:24:18 crc kubenswrapper[4806]: I1204 04:24:18.423762 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:24:18 crc kubenswrapper[4806]: E1204 04:24:18.424425 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:24:19 crc kubenswrapper[4806]: I1204 04:24:19.438620 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="277e6177-7249-4994-9228-c697d5eecff1" path="/var/lib/kubelet/pods/277e6177-7249-4994-9228-c697d5eecff1/volumes" Dec 04 04:24:30 crc kubenswrapper[4806]: I1204 04:24:30.423674 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:24:30 crc kubenswrapper[4806]: I1204 04:24:30.714966 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"c15d4fe56702198504fcf9c93df44bc149f0a2bbdf70116fa6a918971b65cecc"} Dec 04 04:24:52 crc kubenswrapper[4806]: I1204 04:24:52.053907 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-klcjk"] Dec 04 04:24:52 crc kubenswrapper[4806]: I1204 04:24:52.061507 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-klcjk"] Dec 04 04:24:53 crc kubenswrapper[4806]: I1204 04:24:53.436722 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8614a49-3244-4947-b65b-bc61a8490e30" path="/var/lib/kubelet/pods/d8614a49-3244-4947-b65b-bc61a8490e30/volumes" Dec 04 04:25:00 crc kubenswrapper[4806]: I1204 04:25:00.183025 4806 scope.go:117] "RemoveContainer" containerID="cc8516e3e42fc34588e518220726fcc7b1333b9f9a4bc5bf79ee92ae4f908035" Dec 04 04:25:00 crc kubenswrapper[4806]: I1204 04:25:00.206238 4806 scope.go:117] "RemoveContainer" containerID="d4415bfd710e982ef75fd954ec910b098cd548cfe8409bdce6c5cd19c6dcb6c7" Dec 04 04:25:00 crc kubenswrapper[4806]: I1204 04:25:00.237915 4806 scope.go:117] "RemoveContainer" containerID="2e6a94ac72e28e8ed436f6a0ed1f91be0e4798d4d5caae9c95969dfa6f5c38e3" Dec 04 04:25:00 crc kubenswrapper[4806]: I1204 04:25:00.287877 4806 scope.go:117] "RemoveContainer" containerID="867d2725e08df6e3291adffe6875b5963c1e8bc21ec7f2ee49a37f07b7f1a8e0" Dec 04 04:25:00 crc kubenswrapper[4806]: I1204 04:25:00.348248 4806 scope.go:117] "RemoveContainer" containerID="311eefa3fc70ca85a9f147992570094cd8822df0a7261165098da44e9736e47f" Dec 04 04:25:03 crc kubenswrapper[4806]: I1204 04:25:03.056965 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-gpzjd"] Dec 04 04:25:03 crc kubenswrapper[4806]: I1204 04:25:03.078320 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-gpzjd"] Dec 04 04:25:03 crc kubenswrapper[4806]: I1204 04:25:03.435041 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8047cee-ef2b-46ab-aca7-faebe1ae8dfc" path="/var/lib/kubelet/pods/a8047cee-ef2b-46ab-aca7-faebe1ae8dfc/volumes" Dec 04 04:25:06 crc kubenswrapper[4806]: I1204 04:25:06.045719 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-sbp9s"] Dec 04 04:25:06 crc kubenswrapper[4806]: I1204 04:25:06.061614 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-sbp9s"] Dec 04 04:25:07 crc kubenswrapper[4806]: I1204 04:25:07.435170 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="babd629a-b0d5-4ba9-9e01-d36f6bc9ac36" path="/var/lib/kubelet/pods/babd629a-b0d5-4ba9-9e01-d36f6bc9ac36/volumes" Dec 04 04:25:20 crc kubenswrapper[4806]: I1204 04:25:20.041998 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-pl5fx"] Dec 04 04:25:20 crc kubenswrapper[4806]: I1204 04:25:20.051006 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-pl5fx"] Dec 04 04:25:21 crc kubenswrapper[4806]: I1204 04:25:21.438871 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1e9cfb9-1741-4a4e-a05d-7d666699876c" path="/var/lib/kubelet/pods/b1e9cfb9-1741-4a4e-a05d-7d666699876c/volumes" Dec 04 04:25:22 crc kubenswrapper[4806]: I1204 04:25:22.035025 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-r76nk"] Dec 04 04:25:22 crc kubenswrapper[4806]: I1204 04:25:22.043191 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-r76nk"] Dec 04 04:25:23 crc kubenswrapper[4806]: I1204 04:25:23.436966 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a539706c-9990-4bf7-b393-6fe0bc9f3b0e" path="/var/lib/kubelet/pods/a539706c-9990-4bf7-b393-6fe0bc9f3b0e/volumes" Dec 04 04:26:00 crc kubenswrapper[4806]: I1204 04:26:00.481173 4806 scope.go:117] "RemoveContainer" containerID="28cf550b5b6b5bf32045b1666c1c389a131dee767915bf17171311afddd538f3" Dec 04 04:26:00 crc kubenswrapper[4806]: I1204 04:26:00.525658 4806 scope.go:117] "RemoveContainer" containerID="19d388c1240f3f615e6c24b37e7ccf33f547b59fa258adddff42e50f7773b7b9" Dec 04 04:26:00 crc kubenswrapper[4806]: I1204 04:26:00.597137 4806 scope.go:117] "RemoveContainer" containerID="e9c5b59d387e80e532bff56638d010b655f50650443eba16c37f1f4c0bef3721" Dec 04 04:26:00 crc kubenswrapper[4806]: I1204 04:26:00.635755 4806 scope.go:117] "RemoveContainer" containerID="af53fa416c57a054383727c3a586fcbe9eae28466916991d55ae07c02d62462c" Dec 04 04:26:14 crc kubenswrapper[4806]: I1204 04:26:14.704781 4806 generic.go:334] "Generic (PLEG): container finished" podID="0b8e6232-8e76-425c-9b02-94e0a3c236cb" containerID="8514593d5c52ceae8d55eb83becaeb2a26e47f33d06cd864ab07d7f6cb1a1cde" exitCode=0 Dec 04 04:26:14 crc kubenswrapper[4806]: I1204 04:26:14.704877 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" event={"ID":"0b8e6232-8e76-425c-9b02-94e0a3c236cb","Type":"ContainerDied","Data":"8514593d5c52ceae8d55eb83becaeb2a26e47f33d06cd864ab07d7f6cb1a1cde"} Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.165368 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.259359 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b8e6232-8e76-425c-9b02-94e0a3c236cb-inventory\") pod \"0b8e6232-8e76-425c-9b02-94e0a3c236cb\" (UID: \"0b8e6232-8e76-425c-9b02-94e0a3c236cb\") " Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.259414 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b8e6232-8e76-425c-9b02-94e0a3c236cb-ssh-key\") pod \"0b8e6232-8e76-425c-9b02-94e0a3c236cb\" (UID: \"0b8e6232-8e76-425c-9b02-94e0a3c236cb\") " Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.259527 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9ll2\" (UniqueName: \"kubernetes.io/projected/0b8e6232-8e76-425c-9b02-94e0a3c236cb-kube-api-access-l9ll2\") pod \"0b8e6232-8e76-425c-9b02-94e0a3c236cb\" (UID: \"0b8e6232-8e76-425c-9b02-94e0a3c236cb\") " Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.271309 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b8e6232-8e76-425c-9b02-94e0a3c236cb-kube-api-access-l9ll2" (OuterVolumeSpecName: "kube-api-access-l9ll2") pod "0b8e6232-8e76-425c-9b02-94e0a3c236cb" (UID: "0b8e6232-8e76-425c-9b02-94e0a3c236cb"). InnerVolumeSpecName "kube-api-access-l9ll2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.291278 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b8e6232-8e76-425c-9b02-94e0a3c236cb-inventory" (OuterVolumeSpecName: "inventory") pod "0b8e6232-8e76-425c-9b02-94e0a3c236cb" (UID: "0b8e6232-8e76-425c-9b02-94e0a3c236cb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.293302 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b8e6232-8e76-425c-9b02-94e0a3c236cb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0b8e6232-8e76-425c-9b02-94e0a3c236cb" (UID: "0b8e6232-8e76-425c-9b02-94e0a3c236cb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.361568 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0b8e6232-8e76-425c-9b02-94e0a3c236cb-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.361620 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0b8e6232-8e76-425c-9b02-94e0a3c236cb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.361633 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9ll2\" (UniqueName: \"kubernetes.io/projected/0b8e6232-8e76-425c-9b02-94e0a3c236cb-kube-api-access-l9ll2\") on node \"crc\" DevicePath \"\"" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.724895 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" event={"ID":"0b8e6232-8e76-425c-9b02-94e0a3c236cb","Type":"ContainerDied","Data":"d14b593f6811ab2946ae9d0b0a3baf528fb441eb309058c8148ed57505019a43"} Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.724944 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d14b593f6811ab2946ae9d0b0a3baf528fb441eb309058c8148ed57505019a43" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.725025 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.872107 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6"] Dec 04 04:26:16 crc kubenswrapper[4806]: E1204 04:26:16.872615 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b8e6232-8e76-425c-9b02-94e0a3c236cb" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.872643 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b8e6232-8e76-425c-9b02-94e0a3c236cb" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.872876 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b8e6232-8e76-425c-9b02-94e0a3c236cb" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.873642 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.877231 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-q8mqk" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.877464 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.877604 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.877765 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.907660 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6"] Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.971710 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6\" (UID: \"bbebb80f-3f15-464c-bb52-ab7dc2a100c4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.971892 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwmqx\" (UniqueName: \"kubernetes.io/projected/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-kube-api-access-zwmqx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6\" (UID: \"bbebb80f-3f15-464c-bb52-ab7dc2a100c4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" Dec 04 04:26:16 crc kubenswrapper[4806]: I1204 04:26:16.971990 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6\" (UID: \"bbebb80f-3f15-464c-bb52-ab7dc2a100c4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" Dec 04 04:26:17 crc kubenswrapper[4806]: I1204 04:26:17.073877 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwmqx\" (UniqueName: \"kubernetes.io/projected/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-kube-api-access-zwmqx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6\" (UID: \"bbebb80f-3f15-464c-bb52-ab7dc2a100c4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" Dec 04 04:26:17 crc kubenswrapper[4806]: I1204 04:26:17.074319 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6\" (UID: \"bbebb80f-3f15-464c-bb52-ab7dc2a100c4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" Dec 04 04:26:17 crc kubenswrapper[4806]: I1204 04:26:17.074449 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6\" (UID: \"bbebb80f-3f15-464c-bb52-ab7dc2a100c4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" Dec 04 04:26:17 crc kubenswrapper[4806]: I1204 04:26:17.080186 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6\" (UID: \"bbebb80f-3f15-464c-bb52-ab7dc2a100c4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" Dec 04 04:26:17 crc kubenswrapper[4806]: I1204 04:26:17.080268 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6\" (UID: \"bbebb80f-3f15-464c-bb52-ab7dc2a100c4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" Dec 04 04:26:17 crc kubenswrapper[4806]: I1204 04:26:17.107029 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwmqx\" (UniqueName: \"kubernetes.io/projected/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-kube-api-access-zwmqx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6\" (UID: \"bbebb80f-3f15-464c-bb52-ab7dc2a100c4\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" Dec 04 04:26:17 crc kubenswrapper[4806]: I1204 04:26:17.192678 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" Dec 04 04:26:17 crc kubenswrapper[4806]: I1204 04:26:17.746315 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6"] Dec 04 04:26:18 crc kubenswrapper[4806]: I1204 04:26:18.741673 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" event={"ID":"bbebb80f-3f15-464c-bb52-ab7dc2a100c4","Type":"ContainerStarted","Data":"eb2f37efe7f22d1e41dca2942d71934c8342563d501a71935b33809359d5b6bd"} Dec 04 04:26:18 crc kubenswrapper[4806]: I1204 04:26:18.742006 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" event={"ID":"bbebb80f-3f15-464c-bb52-ab7dc2a100c4","Type":"ContainerStarted","Data":"50b5cafedec0a7a9be189a4371bee639832a66859023092cec983415c46a753c"} Dec 04 04:26:18 crc kubenswrapper[4806]: I1204 04:26:18.762569 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" podStartSLOduration=2.482475303 podStartE2EDuration="2.762549585s" podCreationTimestamp="2025-12-04 04:26:16 +0000 UTC" firstStartedPulling="2025-12-04 04:26:17.757524795 +0000 UTC m=+1892.616037723" lastFinishedPulling="2025-12-04 04:26:18.037599077 +0000 UTC m=+1892.896112005" observedRunningTime="2025-12-04 04:26:18.756304389 +0000 UTC m=+1893.614817317" watchObservedRunningTime="2025-12-04 04:26:18.762549585 +0000 UTC m=+1893.621062533" Dec 04 04:26:19 crc kubenswrapper[4806]: I1204 04:26:19.041199 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-61a2-account-create-update-h59lb"] Dec 04 04:26:19 crc kubenswrapper[4806]: I1204 04:26:19.048780 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-61a2-account-create-update-h59lb"] Dec 04 04:26:19 crc kubenswrapper[4806]: I1204 04:26:19.435269 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86cdc393-070d-423a-869d-bf2c9f5be5f7" path="/var/lib/kubelet/pods/86cdc393-070d-423a-869d-bf2c9f5be5f7/volumes" Dec 04 04:26:20 crc kubenswrapper[4806]: I1204 04:26:20.036914 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-72b9-account-create-update-pqj5w"] Dec 04 04:26:20 crc kubenswrapper[4806]: I1204 04:26:20.047977 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-vpg5m"] Dec 04 04:26:20 crc kubenswrapper[4806]: I1204 04:26:20.060422 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-91c0-account-create-update-2fbzf"] Dec 04 04:26:20 crc kubenswrapper[4806]: I1204 04:26:20.072359 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-wqwpc"] Dec 04 04:26:20 crc kubenswrapper[4806]: I1204 04:26:20.083195 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-jt65t"] Dec 04 04:26:20 crc kubenswrapper[4806]: I1204 04:26:20.094694 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-72b9-account-create-update-pqj5w"] Dec 04 04:26:20 crc kubenswrapper[4806]: I1204 04:26:20.106500 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-vpg5m"] Dec 04 04:26:20 crc kubenswrapper[4806]: I1204 04:26:20.114042 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-wqwpc"] Dec 04 04:26:20 crc kubenswrapper[4806]: I1204 04:26:20.123017 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-91c0-account-create-update-2fbzf"] Dec 04 04:26:20 crc kubenswrapper[4806]: I1204 04:26:20.132318 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-jt65t"] Dec 04 04:26:21 crc kubenswrapper[4806]: I1204 04:26:21.436147 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10d9b83f-5515-45d6-86c7-3331dad07c2e" path="/var/lib/kubelet/pods/10d9b83f-5515-45d6-86c7-3331dad07c2e/volumes" Dec 04 04:26:21 crc kubenswrapper[4806]: I1204 04:26:21.438022 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="158f847a-66f8-42c6-b4e0-242dad5a702c" path="/var/lib/kubelet/pods/158f847a-66f8-42c6-b4e0-242dad5a702c/volumes" Dec 04 04:26:21 crc kubenswrapper[4806]: I1204 04:26:21.439271 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37d77210-b359-46a2-879c-5422946b6341" path="/var/lib/kubelet/pods/37d77210-b359-46a2-879c-5422946b6341/volumes" Dec 04 04:26:21 crc kubenswrapper[4806]: I1204 04:26:21.440887 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38dd0901-0672-4cac-93d0-881608300b55" path="/var/lib/kubelet/pods/38dd0901-0672-4cac-93d0-881608300b55/volumes" Dec 04 04:26:21 crc kubenswrapper[4806]: I1204 04:26:21.442883 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da637fcd-562f-4e71-a068-a4f74b94663f" path="/var/lib/kubelet/pods/da637fcd-562f-4e71-a068-a4f74b94663f/volumes" Dec 04 04:26:54 crc kubenswrapper[4806]: I1204 04:26:54.042440 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xcbb8"] Dec 04 04:26:54 crc kubenswrapper[4806]: I1204 04:26:54.057108 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xcbb8"] Dec 04 04:26:55 crc kubenswrapper[4806]: I1204 04:26:55.433941 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f577bb40-09d5-43ac-ae7b-1e587eed300e" path="/var/lib/kubelet/pods/f577bb40-09d5-43ac-ae7b-1e587eed300e/volumes" Dec 04 04:26:57 crc kubenswrapper[4806]: I1204 04:26:57.047238 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:26:57 crc kubenswrapper[4806]: I1204 04:26:57.048325 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:27:00 crc kubenswrapper[4806]: I1204 04:27:00.750365 4806 scope.go:117] "RemoveContainer" containerID="bfadb08507dee6b7a2cb7b6f098d0f50f0840a50e4f3ce650b4728c94302a898" Dec 04 04:27:00 crc kubenswrapper[4806]: I1204 04:27:00.779063 4806 scope.go:117] "RemoveContainer" containerID="c69e2e3bc7ed29c334fb3b328c3e2b97610be6079558e730d5a27943dff668e7" Dec 04 04:27:00 crc kubenswrapper[4806]: I1204 04:27:00.857483 4806 scope.go:117] "RemoveContainer" containerID="193f0517c4150de6b45ddc9f5267b7220800b401d23ac394dbe24d3b415ab5e5" Dec 04 04:27:00 crc kubenswrapper[4806]: I1204 04:27:00.911147 4806 scope.go:117] "RemoveContainer" containerID="313f402869fe32a6b803febc507ed2f07322acc2418edf9a318d34579cd1eb0c" Dec 04 04:27:01 crc kubenswrapper[4806]: I1204 04:27:01.003041 4806 scope.go:117] "RemoveContainer" containerID="8d6d9ec2c5ccf12239923ee0c0db74aa93fe75dbbaa6b014b81b0a2ab620c83a" Dec 04 04:27:01 crc kubenswrapper[4806]: I1204 04:27:01.028563 4806 scope.go:117] "RemoveContainer" containerID="02f52c56412c0e77c78a513b111c8085b365bc749210a2a9e0701d217747edfd" Dec 04 04:27:01 crc kubenswrapper[4806]: I1204 04:27:01.074080 4806 scope.go:117] "RemoveContainer" containerID="91568e3250f24fb58e5225f894168d86d33337304aa52036d0ace330822e95a6" Dec 04 04:27:21 crc kubenswrapper[4806]: I1204 04:27:21.059489 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-zrdpr"] Dec 04 04:27:21 crc kubenswrapper[4806]: I1204 04:27:21.070798 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-zrdpr"] Dec 04 04:27:21 crc kubenswrapper[4806]: I1204 04:27:21.452036 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79a4caa8-a6ad-47a0-8562-9217d129b010" path="/var/lib/kubelet/pods/79a4caa8-a6ad-47a0-8562-9217d129b010/volumes" Dec 04 04:27:23 crc kubenswrapper[4806]: I1204 04:27:23.031253 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7h76s"] Dec 04 04:27:23 crc kubenswrapper[4806]: I1204 04:27:23.040629 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7h76s"] Dec 04 04:27:23 crc kubenswrapper[4806]: I1204 04:27:23.435083 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec8caa98-4401-4f0b-86fa-ff3d04e00544" path="/var/lib/kubelet/pods/ec8caa98-4401-4f0b-86fa-ff3d04e00544/volumes" Dec 04 04:27:27 crc kubenswrapper[4806]: I1204 04:27:27.046877 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:27:27 crc kubenswrapper[4806]: I1204 04:27:27.047488 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:27:42 crc kubenswrapper[4806]: I1204 04:27:42.819359 4806 generic.go:334] "Generic (PLEG): container finished" podID="bbebb80f-3f15-464c-bb52-ab7dc2a100c4" containerID="eb2f37efe7f22d1e41dca2942d71934c8342563d501a71935b33809359d5b6bd" exitCode=0 Dec 04 04:27:42 crc kubenswrapper[4806]: I1204 04:27:42.819458 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" event={"ID":"bbebb80f-3f15-464c-bb52-ab7dc2a100c4","Type":"ContainerDied","Data":"eb2f37efe7f22d1e41dca2942d71934c8342563d501a71935b33809359d5b6bd"} Dec 04 04:27:44 crc kubenswrapper[4806]: I1204 04:27:44.250463 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" Dec 04 04:27:44 crc kubenswrapper[4806]: I1204 04:27:44.352628 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-ssh-key\") pod \"bbebb80f-3f15-464c-bb52-ab7dc2a100c4\" (UID: \"bbebb80f-3f15-464c-bb52-ab7dc2a100c4\") " Dec 04 04:27:44 crc kubenswrapper[4806]: I1204 04:27:44.352684 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-inventory\") pod \"bbebb80f-3f15-464c-bb52-ab7dc2a100c4\" (UID: \"bbebb80f-3f15-464c-bb52-ab7dc2a100c4\") " Dec 04 04:27:44 crc kubenswrapper[4806]: I1204 04:27:44.352842 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwmqx\" (UniqueName: \"kubernetes.io/projected/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-kube-api-access-zwmqx\") pod \"bbebb80f-3f15-464c-bb52-ab7dc2a100c4\" (UID: \"bbebb80f-3f15-464c-bb52-ab7dc2a100c4\") " Dec 04 04:27:44 crc kubenswrapper[4806]: I1204 04:27:44.364132 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-kube-api-access-zwmqx" (OuterVolumeSpecName: "kube-api-access-zwmqx") pod "bbebb80f-3f15-464c-bb52-ab7dc2a100c4" (UID: "bbebb80f-3f15-464c-bb52-ab7dc2a100c4"). InnerVolumeSpecName "kube-api-access-zwmqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:27:44 crc kubenswrapper[4806]: I1204 04:27:44.380689 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-inventory" (OuterVolumeSpecName: "inventory") pod "bbebb80f-3f15-464c-bb52-ab7dc2a100c4" (UID: "bbebb80f-3f15-464c-bb52-ab7dc2a100c4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:27:44 crc kubenswrapper[4806]: I1204 04:27:44.386060 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bbebb80f-3f15-464c-bb52-ab7dc2a100c4" (UID: "bbebb80f-3f15-464c-bb52-ab7dc2a100c4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:27:44 crc kubenswrapper[4806]: I1204 04:27:44.454992 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:27:44 crc kubenswrapper[4806]: I1204 04:27:44.455023 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 04:27:44 crc kubenswrapper[4806]: I1204 04:27:44.455033 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwmqx\" (UniqueName: \"kubernetes.io/projected/bbebb80f-3f15-464c-bb52-ab7dc2a100c4-kube-api-access-zwmqx\") on node \"crc\" DevicePath \"\"" Dec 04 04:27:44 crc kubenswrapper[4806]: I1204 04:27:44.842854 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" event={"ID":"bbebb80f-3f15-464c-bb52-ab7dc2a100c4","Type":"ContainerDied","Data":"50b5cafedec0a7a9be189a4371bee639832a66859023092cec983415c46a753c"} Dec 04 04:27:44 crc kubenswrapper[4806]: I1204 04:27:44.843103 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50b5cafedec0a7a9be189a4371bee639832a66859023092cec983415c46a753c" Dec 04 04:27:44 crc kubenswrapper[4806]: I1204 04:27:44.842901 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.004790 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h"] Dec 04 04:27:45 crc kubenswrapper[4806]: E1204 04:27:45.005410 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbebb80f-3f15-464c-bb52-ab7dc2a100c4" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.005432 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbebb80f-3f15-464c-bb52-ab7dc2a100c4" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.005702 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbebb80f-3f15-464c-bb52-ab7dc2a100c4" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.006480 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.013501 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-q8mqk" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.013809 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.013999 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.013667 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.018094 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h"] Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.066501 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02021c61-fa0c-4aea-80d4-b0eb592c0d06-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phq5h\" (UID: \"02021c61-fa0c-4aea-80d4-b0eb592c0d06\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.066585 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02021c61-fa0c-4aea-80d4-b0eb592c0d06-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phq5h\" (UID: \"02021c61-fa0c-4aea-80d4-b0eb592c0d06\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.066733 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsx8l\" (UniqueName: \"kubernetes.io/projected/02021c61-fa0c-4aea-80d4-b0eb592c0d06-kube-api-access-rsx8l\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phq5h\" (UID: \"02021c61-fa0c-4aea-80d4-b0eb592c0d06\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.168555 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02021c61-fa0c-4aea-80d4-b0eb592c0d06-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phq5h\" (UID: \"02021c61-fa0c-4aea-80d4-b0eb592c0d06\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.169042 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsx8l\" (UniqueName: \"kubernetes.io/projected/02021c61-fa0c-4aea-80d4-b0eb592c0d06-kube-api-access-rsx8l\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phq5h\" (UID: \"02021c61-fa0c-4aea-80d4-b0eb592c0d06\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.169222 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02021c61-fa0c-4aea-80d4-b0eb592c0d06-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phq5h\" (UID: \"02021c61-fa0c-4aea-80d4-b0eb592c0d06\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.173572 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02021c61-fa0c-4aea-80d4-b0eb592c0d06-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phq5h\" (UID: \"02021c61-fa0c-4aea-80d4-b0eb592c0d06\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.173958 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02021c61-fa0c-4aea-80d4-b0eb592c0d06-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phq5h\" (UID: \"02021c61-fa0c-4aea-80d4-b0eb592c0d06\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.187752 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsx8l\" (UniqueName: \"kubernetes.io/projected/02021c61-fa0c-4aea-80d4-b0eb592c0d06-kube-api-access-rsx8l\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-phq5h\" (UID: \"02021c61-fa0c-4aea-80d4-b0eb592c0d06\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.327258 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.888718 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h"] Dec 04 04:27:45 crc kubenswrapper[4806]: I1204 04:27:45.908870 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 04:27:46 crc kubenswrapper[4806]: I1204 04:27:46.135953 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:27:46 crc kubenswrapper[4806]: I1204 04:27:46.864873 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" event={"ID":"02021c61-fa0c-4aea-80d4-b0eb592c0d06","Type":"ContainerStarted","Data":"6d92194d485e5cda4b5da40d835bd9f22ffbc34cf4aee9922b938c14ac39069d"} Dec 04 04:27:46 crc kubenswrapper[4806]: I1204 04:27:46.865201 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" event={"ID":"02021c61-fa0c-4aea-80d4-b0eb592c0d06","Type":"ContainerStarted","Data":"de06149fd407d92fb9730ad6111cf308145dab876c6c5d4857450ffad68aa670"} Dec 04 04:27:46 crc kubenswrapper[4806]: I1204 04:27:46.888444 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" podStartSLOduration=2.66622583 podStartE2EDuration="2.888424806s" podCreationTimestamp="2025-12-04 04:27:44 +0000 UTC" firstStartedPulling="2025-12-04 04:27:45.908652474 +0000 UTC m=+1980.767165402" lastFinishedPulling="2025-12-04 04:27:46.13085145 +0000 UTC m=+1980.989364378" observedRunningTime="2025-12-04 04:27:46.882759029 +0000 UTC m=+1981.741271967" watchObservedRunningTime="2025-12-04 04:27:46.888424806 +0000 UTC m=+1981.746937734" Dec 04 04:27:51 crc kubenswrapper[4806]: I1204 04:27:51.924079 4806 generic.go:334] "Generic (PLEG): container finished" podID="02021c61-fa0c-4aea-80d4-b0eb592c0d06" containerID="6d92194d485e5cda4b5da40d835bd9f22ffbc34cf4aee9922b938c14ac39069d" exitCode=0 Dec 04 04:27:51 crc kubenswrapper[4806]: I1204 04:27:51.924664 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" event={"ID":"02021c61-fa0c-4aea-80d4-b0eb592c0d06","Type":"ContainerDied","Data":"6d92194d485e5cda4b5da40d835bd9f22ffbc34cf4aee9922b938c14ac39069d"} Dec 04 04:27:53 crc kubenswrapper[4806]: I1204 04:27:53.690661 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" Dec 04 04:27:53 crc kubenswrapper[4806]: I1204 04:27:53.887490 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsx8l\" (UniqueName: \"kubernetes.io/projected/02021c61-fa0c-4aea-80d4-b0eb592c0d06-kube-api-access-rsx8l\") pod \"02021c61-fa0c-4aea-80d4-b0eb592c0d06\" (UID: \"02021c61-fa0c-4aea-80d4-b0eb592c0d06\") " Dec 04 04:27:53 crc kubenswrapper[4806]: I1204 04:27:53.887567 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02021c61-fa0c-4aea-80d4-b0eb592c0d06-ssh-key\") pod \"02021c61-fa0c-4aea-80d4-b0eb592c0d06\" (UID: \"02021c61-fa0c-4aea-80d4-b0eb592c0d06\") " Dec 04 04:27:53 crc kubenswrapper[4806]: I1204 04:27:53.887611 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02021c61-fa0c-4aea-80d4-b0eb592c0d06-inventory\") pod \"02021c61-fa0c-4aea-80d4-b0eb592c0d06\" (UID: \"02021c61-fa0c-4aea-80d4-b0eb592c0d06\") " Dec 04 04:27:53 crc kubenswrapper[4806]: I1204 04:27:53.898350 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02021c61-fa0c-4aea-80d4-b0eb592c0d06-kube-api-access-rsx8l" (OuterVolumeSpecName: "kube-api-access-rsx8l") pod "02021c61-fa0c-4aea-80d4-b0eb592c0d06" (UID: "02021c61-fa0c-4aea-80d4-b0eb592c0d06"). InnerVolumeSpecName "kube-api-access-rsx8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:27:53 crc kubenswrapper[4806]: I1204 04:27:53.917701 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02021c61-fa0c-4aea-80d4-b0eb592c0d06-inventory" (OuterVolumeSpecName: "inventory") pod "02021c61-fa0c-4aea-80d4-b0eb592c0d06" (UID: "02021c61-fa0c-4aea-80d4-b0eb592c0d06"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:27:53 crc kubenswrapper[4806]: I1204 04:27:53.922319 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02021c61-fa0c-4aea-80d4-b0eb592c0d06-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "02021c61-fa0c-4aea-80d4-b0eb592c0d06" (UID: "02021c61-fa0c-4aea-80d4-b0eb592c0d06"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:27:53 crc kubenswrapper[4806]: I1204 04:27:53.945867 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" event={"ID":"02021c61-fa0c-4aea-80d4-b0eb592c0d06","Type":"ContainerDied","Data":"de06149fd407d92fb9730ad6111cf308145dab876c6c5d4857450ffad68aa670"} Dec 04 04:27:53 crc kubenswrapper[4806]: I1204 04:27:53.945960 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de06149fd407d92fb9730ad6111cf308145dab876c6c5d4857450ffad68aa670" Dec 04 04:27:53 crc kubenswrapper[4806]: I1204 04:27:53.945982 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-phq5h" Dec 04 04:27:53 crc kubenswrapper[4806]: I1204 04:27:53.991196 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsx8l\" (UniqueName: \"kubernetes.io/projected/02021c61-fa0c-4aea-80d4-b0eb592c0d06-kube-api-access-rsx8l\") on node \"crc\" DevicePath \"\"" Dec 04 04:27:53 crc kubenswrapper[4806]: I1204 04:27:53.991279 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02021c61-fa0c-4aea-80d4-b0eb592c0d06-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:27:53 crc kubenswrapper[4806]: I1204 04:27:53.991299 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02021c61-fa0c-4aea-80d4-b0eb592c0d06-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.019117 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25"] Dec 04 04:27:54 crc kubenswrapper[4806]: E1204 04:27:54.019872 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02021c61-fa0c-4aea-80d4-b0eb592c0d06" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.019989 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="02021c61-fa0c-4aea-80d4-b0eb592c0d06" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.020356 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="02021c61-fa0c-4aea-80d4-b0eb592c0d06" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.021135 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.025345 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.026028 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-q8mqk" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.026198 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.027241 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.081618 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25"] Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.198691 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a53f8ff-b197-4f2b-8194-6424a21dae22-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fdr25\" (UID: \"3a53f8ff-b197-4f2b-8194-6424a21dae22\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.198954 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvfpg\" (UniqueName: \"kubernetes.io/projected/3a53f8ff-b197-4f2b-8194-6424a21dae22-kube-api-access-bvfpg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fdr25\" (UID: \"3a53f8ff-b197-4f2b-8194-6424a21dae22\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.199044 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a53f8ff-b197-4f2b-8194-6424a21dae22-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fdr25\" (UID: \"3a53f8ff-b197-4f2b-8194-6424a21dae22\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.301710 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvfpg\" (UniqueName: \"kubernetes.io/projected/3a53f8ff-b197-4f2b-8194-6424a21dae22-kube-api-access-bvfpg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fdr25\" (UID: \"3a53f8ff-b197-4f2b-8194-6424a21dae22\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.301794 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a53f8ff-b197-4f2b-8194-6424a21dae22-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fdr25\" (UID: \"3a53f8ff-b197-4f2b-8194-6424a21dae22\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.302004 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a53f8ff-b197-4f2b-8194-6424a21dae22-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fdr25\" (UID: \"3a53f8ff-b197-4f2b-8194-6424a21dae22\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.310056 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a53f8ff-b197-4f2b-8194-6424a21dae22-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fdr25\" (UID: \"3a53f8ff-b197-4f2b-8194-6424a21dae22\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.310151 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a53f8ff-b197-4f2b-8194-6424a21dae22-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fdr25\" (UID: \"3a53f8ff-b197-4f2b-8194-6424a21dae22\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.328979 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvfpg\" (UniqueName: \"kubernetes.io/projected/3a53f8ff-b197-4f2b-8194-6424a21dae22-kube-api-access-bvfpg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fdr25\" (UID: \"3a53f8ff-b197-4f2b-8194-6424a21dae22\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" Dec 04 04:27:54 crc kubenswrapper[4806]: I1204 04:27:54.374001 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" Dec 04 04:27:55 crc kubenswrapper[4806]: I1204 04:27:55.002458 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25"] Dec 04 04:27:55 crc kubenswrapper[4806]: W1204 04:27:55.013858 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a53f8ff_b197_4f2b_8194_6424a21dae22.slice/crio-9dd7eb766dd0767ba7b997f40700432094a43aeb4d79ef63ec284fb4e20e5147 WatchSource:0}: Error finding container 9dd7eb766dd0767ba7b997f40700432094a43aeb4d79ef63ec284fb4e20e5147: Status 404 returned error can't find the container with id 9dd7eb766dd0767ba7b997f40700432094a43aeb4d79ef63ec284fb4e20e5147 Dec 04 04:27:55 crc kubenswrapper[4806]: I1204 04:27:55.980602 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" event={"ID":"3a53f8ff-b197-4f2b-8194-6424a21dae22","Type":"ContainerStarted","Data":"214d3929c526297da391bf70caccbef0a3a474b5147f77ca98a4d39c5c9f946c"} Dec 04 04:27:55 crc kubenswrapper[4806]: I1204 04:27:55.981062 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" event={"ID":"3a53f8ff-b197-4f2b-8194-6424a21dae22","Type":"ContainerStarted","Data":"9dd7eb766dd0767ba7b997f40700432094a43aeb4d79ef63ec284fb4e20e5147"} Dec 04 04:27:56 crc kubenswrapper[4806]: I1204 04:27:56.013034 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" podStartSLOduration=2.814112725 podStartE2EDuration="3.012997632s" podCreationTimestamp="2025-12-04 04:27:53 +0000 UTC" firstStartedPulling="2025-12-04 04:27:55.020847124 +0000 UTC m=+1989.879360052" lastFinishedPulling="2025-12-04 04:27:55.219732031 +0000 UTC m=+1990.078244959" observedRunningTime="2025-12-04 04:27:56.003772444 +0000 UTC m=+1990.862285382" watchObservedRunningTime="2025-12-04 04:27:56.012997632 +0000 UTC m=+1990.871510560" Dec 04 04:27:57 crc kubenswrapper[4806]: I1204 04:27:57.048204 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:27:57 crc kubenswrapper[4806]: I1204 04:27:57.048836 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:27:57 crc kubenswrapper[4806]: I1204 04:27:57.048959 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 04:27:57 crc kubenswrapper[4806]: I1204 04:27:57.050251 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c15d4fe56702198504fcf9c93df44bc149f0a2bbdf70116fa6a918971b65cecc"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 04:27:57 crc kubenswrapper[4806]: I1204 04:27:57.050331 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://c15d4fe56702198504fcf9c93df44bc149f0a2bbdf70116fa6a918971b65cecc" gracePeriod=600 Dec 04 04:27:58 crc kubenswrapper[4806]: I1204 04:27:58.031685 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="c15d4fe56702198504fcf9c93df44bc149f0a2bbdf70116fa6a918971b65cecc" exitCode=0 Dec 04 04:27:58 crc kubenswrapper[4806]: I1204 04:27:58.032089 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"c15d4fe56702198504fcf9c93df44bc149f0a2bbdf70116fa6a918971b65cecc"} Dec 04 04:27:58 crc kubenswrapper[4806]: I1204 04:27:58.032245 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513"} Dec 04 04:27:58 crc kubenswrapper[4806]: I1204 04:27:58.032292 4806 scope.go:117] "RemoveContainer" containerID="552de1d68efdc569fead29a3a8ba003f6cef1b3c5161128527c6965edd11e6bc" Dec 04 04:28:01 crc kubenswrapper[4806]: I1204 04:28:01.189045 4806 scope.go:117] "RemoveContainer" containerID="060ce8d81fcce975a2212d682de78807111fc2db24634bdc8b95148fa2681ed4" Dec 04 04:28:01 crc kubenswrapper[4806]: I1204 04:28:01.230739 4806 scope.go:117] "RemoveContainer" containerID="9981c8f74b5875c56bf20e832a666f1ec7a31accd5a8839c5c0c0b3146dda24c" Dec 04 04:28:08 crc kubenswrapper[4806]: I1204 04:28:08.048188 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-fgv9w"] Dec 04 04:28:08 crc kubenswrapper[4806]: I1204 04:28:08.059439 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-fgv9w"] Dec 04 04:28:09 crc kubenswrapper[4806]: I1204 04:28:09.436644 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1f08876-fc00-4755-82ca-87352049c5d3" path="/var/lib/kubelet/pods/d1f08876-fc00-4755-82ca-87352049c5d3/volumes" Dec 04 04:28:36 crc kubenswrapper[4806]: I1204 04:28:36.417383 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dtdft"] Dec 04 04:28:36 crc kubenswrapper[4806]: I1204 04:28:36.420444 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:36 crc kubenswrapper[4806]: I1204 04:28:36.434678 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dtdft"] Dec 04 04:28:36 crc kubenswrapper[4806]: I1204 04:28:36.476567 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c047d3a7-46e0-4f42-ac87-22700753f4f1-catalog-content\") pod \"redhat-operators-dtdft\" (UID: \"c047d3a7-46e0-4f42-ac87-22700753f4f1\") " pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:36 crc kubenswrapper[4806]: I1204 04:28:36.476646 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckl8w\" (UniqueName: \"kubernetes.io/projected/c047d3a7-46e0-4f42-ac87-22700753f4f1-kube-api-access-ckl8w\") pod \"redhat-operators-dtdft\" (UID: \"c047d3a7-46e0-4f42-ac87-22700753f4f1\") " pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:36 crc kubenswrapper[4806]: I1204 04:28:36.476712 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c047d3a7-46e0-4f42-ac87-22700753f4f1-utilities\") pod \"redhat-operators-dtdft\" (UID: \"c047d3a7-46e0-4f42-ac87-22700753f4f1\") " pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:36 crc kubenswrapper[4806]: I1204 04:28:36.578857 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c047d3a7-46e0-4f42-ac87-22700753f4f1-catalog-content\") pod \"redhat-operators-dtdft\" (UID: \"c047d3a7-46e0-4f42-ac87-22700753f4f1\") " pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:36 crc kubenswrapper[4806]: I1204 04:28:36.578967 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckl8w\" (UniqueName: \"kubernetes.io/projected/c047d3a7-46e0-4f42-ac87-22700753f4f1-kube-api-access-ckl8w\") pod \"redhat-operators-dtdft\" (UID: \"c047d3a7-46e0-4f42-ac87-22700753f4f1\") " pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:36 crc kubenswrapper[4806]: I1204 04:28:36.579024 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c047d3a7-46e0-4f42-ac87-22700753f4f1-utilities\") pod \"redhat-operators-dtdft\" (UID: \"c047d3a7-46e0-4f42-ac87-22700753f4f1\") " pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:36 crc kubenswrapper[4806]: I1204 04:28:36.579487 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c047d3a7-46e0-4f42-ac87-22700753f4f1-catalog-content\") pod \"redhat-operators-dtdft\" (UID: \"c047d3a7-46e0-4f42-ac87-22700753f4f1\") " pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:36 crc kubenswrapper[4806]: I1204 04:28:36.579577 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c047d3a7-46e0-4f42-ac87-22700753f4f1-utilities\") pod \"redhat-operators-dtdft\" (UID: \"c047d3a7-46e0-4f42-ac87-22700753f4f1\") " pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:36 crc kubenswrapper[4806]: I1204 04:28:36.612453 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckl8w\" (UniqueName: \"kubernetes.io/projected/c047d3a7-46e0-4f42-ac87-22700753f4f1-kube-api-access-ckl8w\") pod \"redhat-operators-dtdft\" (UID: \"c047d3a7-46e0-4f42-ac87-22700753f4f1\") " pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:36 crc kubenswrapper[4806]: I1204 04:28:36.742186 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:37 crc kubenswrapper[4806]: I1204 04:28:37.315842 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dtdft"] Dec 04 04:28:37 crc kubenswrapper[4806]: I1204 04:28:37.458662 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtdft" event={"ID":"c047d3a7-46e0-4f42-ac87-22700753f4f1","Type":"ContainerStarted","Data":"1b179a058a3aab809f48f6df3bd66379f91d30ce29c3d140aea3f9df42e48214"} Dec 04 04:28:38 crc kubenswrapper[4806]: I1204 04:28:38.469108 4806 generic.go:334] "Generic (PLEG): container finished" podID="c047d3a7-46e0-4f42-ac87-22700753f4f1" containerID="21f7e6be351bd95dd978d8e2b08f5538e004755eb96bc2cbe9132a5452b59af4" exitCode=0 Dec 04 04:28:38 crc kubenswrapper[4806]: I1204 04:28:38.469198 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtdft" event={"ID":"c047d3a7-46e0-4f42-ac87-22700753f4f1","Type":"ContainerDied","Data":"21f7e6be351bd95dd978d8e2b08f5538e004755eb96bc2cbe9132a5452b59af4"} Dec 04 04:28:39 crc kubenswrapper[4806]: I1204 04:28:39.485526 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtdft" event={"ID":"c047d3a7-46e0-4f42-ac87-22700753f4f1","Type":"ContainerStarted","Data":"63b98617e04d9a60b350806cc52d29e825df6188083d82c626598a8b0785b837"} Dec 04 04:28:41 crc kubenswrapper[4806]: I1204 04:28:41.505794 4806 generic.go:334] "Generic (PLEG): container finished" podID="3a53f8ff-b197-4f2b-8194-6424a21dae22" containerID="214d3929c526297da391bf70caccbef0a3a474b5147f77ca98a4d39c5c9f946c" exitCode=0 Dec 04 04:28:41 crc kubenswrapper[4806]: I1204 04:28:41.505879 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" event={"ID":"3a53f8ff-b197-4f2b-8194-6424a21dae22","Type":"ContainerDied","Data":"214d3929c526297da391bf70caccbef0a3a474b5147f77ca98a4d39c5c9f946c"} Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.049173 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.221771 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvfpg\" (UniqueName: \"kubernetes.io/projected/3a53f8ff-b197-4f2b-8194-6424a21dae22-kube-api-access-bvfpg\") pod \"3a53f8ff-b197-4f2b-8194-6424a21dae22\" (UID: \"3a53f8ff-b197-4f2b-8194-6424a21dae22\") " Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.222183 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a53f8ff-b197-4f2b-8194-6424a21dae22-inventory\") pod \"3a53f8ff-b197-4f2b-8194-6424a21dae22\" (UID: \"3a53f8ff-b197-4f2b-8194-6424a21dae22\") " Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.222364 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a53f8ff-b197-4f2b-8194-6424a21dae22-ssh-key\") pod \"3a53f8ff-b197-4f2b-8194-6424a21dae22\" (UID: \"3a53f8ff-b197-4f2b-8194-6424a21dae22\") " Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.228362 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a53f8ff-b197-4f2b-8194-6424a21dae22-kube-api-access-bvfpg" (OuterVolumeSpecName: "kube-api-access-bvfpg") pod "3a53f8ff-b197-4f2b-8194-6424a21dae22" (UID: "3a53f8ff-b197-4f2b-8194-6424a21dae22"). InnerVolumeSpecName "kube-api-access-bvfpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.248810 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a53f8ff-b197-4f2b-8194-6424a21dae22-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3a53f8ff-b197-4f2b-8194-6424a21dae22" (UID: "3a53f8ff-b197-4f2b-8194-6424a21dae22"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.254811 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a53f8ff-b197-4f2b-8194-6424a21dae22-inventory" (OuterVolumeSpecName: "inventory") pod "3a53f8ff-b197-4f2b-8194-6424a21dae22" (UID: "3a53f8ff-b197-4f2b-8194-6424a21dae22"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.324610 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3a53f8ff-b197-4f2b-8194-6424a21dae22-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.324653 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvfpg\" (UniqueName: \"kubernetes.io/projected/3a53f8ff-b197-4f2b-8194-6424a21dae22-kube-api-access-bvfpg\") on node \"crc\" DevicePath \"\"" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.324668 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3a53f8ff-b197-4f2b-8194-6424a21dae22-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.530678 4806 generic.go:334] "Generic (PLEG): container finished" podID="c047d3a7-46e0-4f42-ac87-22700753f4f1" containerID="63b98617e04d9a60b350806cc52d29e825df6188083d82c626598a8b0785b837" exitCode=0 Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.530752 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtdft" event={"ID":"c047d3a7-46e0-4f42-ac87-22700753f4f1","Type":"ContainerDied","Data":"63b98617e04d9a60b350806cc52d29e825df6188083d82c626598a8b0785b837"} Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.533131 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" event={"ID":"3a53f8ff-b197-4f2b-8194-6424a21dae22","Type":"ContainerDied","Data":"9dd7eb766dd0767ba7b997f40700432094a43aeb4d79ef63ec284fb4e20e5147"} Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.533193 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9dd7eb766dd0767ba7b997f40700432094a43aeb4d79ef63ec284fb4e20e5147" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.533288 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fdr25" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.639963 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8"] Dec 04 04:28:43 crc kubenswrapper[4806]: E1204 04:28:43.640463 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a53f8ff-b197-4f2b-8194-6424a21dae22" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.640488 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a53f8ff-b197-4f2b-8194-6424a21dae22" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.640746 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a53f8ff-b197-4f2b-8194-6424a21dae22" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.641491 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.644641 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.644805 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.645292 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.646445 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-q8mqk" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.662729 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8"] Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.733327 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-984f8\" (UID: \"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.733493 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf4k5\" (UniqueName: \"kubernetes.io/projected/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-kube-api-access-sf4k5\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-984f8\" (UID: \"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.733548 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-984f8\" (UID: \"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.835064 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-984f8\" (UID: \"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.835169 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-984f8\" (UID: \"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.835462 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf4k5\" (UniqueName: \"kubernetes.io/projected/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-kube-api-access-sf4k5\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-984f8\" (UID: \"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.839385 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-984f8\" (UID: \"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.845516 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-984f8\" (UID: \"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.853703 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf4k5\" (UniqueName: \"kubernetes.io/projected/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-kube-api-access-sf4k5\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-984f8\" (UID: \"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" Dec 04 04:28:43 crc kubenswrapper[4806]: I1204 04:28:43.960608 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" Dec 04 04:28:44 crc kubenswrapper[4806]: I1204 04:28:44.535946 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8"] Dec 04 04:28:44 crc kubenswrapper[4806]: I1204 04:28:44.546632 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" event={"ID":"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c","Type":"ContainerStarted","Data":"73d892f6d29c122f5e7e16999ff2d9e40362b6e5427120a40ff2d8f9534e71bb"} Dec 04 04:28:44 crc kubenswrapper[4806]: I1204 04:28:44.549183 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtdft" event={"ID":"c047d3a7-46e0-4f42-ac87-22700753f4f1","Type":"ContainerStarted","Data":"f94c31ed94d9c481de0eb00423963c7b57d62b878b592b3d60cbca09d808d41f"} Dec 04 04:28:44 crc kubenswrapper[4806]: I1204 04:28:44.572608 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dtdft" podStartSLOduration=3.083470066 podStartE2EDuration="8.572588377s" podCreationTimestamp="2025-12-04 04:28:36 +0000 UTC" firstStartedPulling="2025-12-04 04:28:38.471969571 +0000 UTC m=+2033.330482519" lastFinishedPulling="2025-12-04 04:28:43.961087902 +0000 UTC m=+2038.819600830" observedRunningTime="2025-12-04 04:28:44.566662592 +0000 UTC m=+2039.425175520" watchObservedRunningTime="2025-12-04 04:28:44.572588377 +0000 UTC m=+2039.431101305" Dec 04 04:28:45 crc kubenswrapper[4806]: I1204 04:28:45.561105 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" event={"ID":"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c","Type":"ContainerStarted","Data":"4ab9c98bf2cac8df8c86fbe81f501b5d312346fffe2b4c3e7b31c005d18be3b0"} Dec 04 04:28:45 crc kubenswrapper[4806]: I1204 04:28:45.592836 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" podStartSLOduration=2.414989511 podStartE2EDuration="2.592807729s" podCreationTimestamp="2025-12-04 04:28:43 +0000 UTC" firstStartedPulling="2025-12-04 04:28:44.538971977 +0000 UTC m=+2039.397484905" lastFinishedPulling="2025-12-04 04:28:44.716790195 +0000 UTC m=+2039.575303123" observedRunningTime="2025-12-04 04:28:45.582496627 +0000 UTC m=+2040.441009565" watchObservedRunningTime="2025-12-04 04:28:45.592807729 +0000 UTC m=+2040.451320657" Dec 04 04:28:46 crc kubenswrapper[4806]: I1204 04:28:46.743175 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:46 crc kubenswrapper[4806]: I1204 04:28:46.743565 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:47 crc kubenswrapper[4806]: I1204 04:28:47.792037 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dtdft" podUID="c047d3a7-46e0-4f42-ac87-22700753f4f1" containerName="registry-server" probeResult="failure" output=< Dec 04 04:28:47 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Dec 04 04:28:47 crc kubenswrapper[4806]: > Dec 04 04:28:56 crc kubenswrapper[4806]: I1204 04:28:56.792728 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:56 crc kubenswrapper[4806]: I1204 04:28:56.861525 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:57 crc kubenswrapper[4806]: I1204 04:28:57.032444 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dtdft"] Dec 04 04:28:58 crc kubenswrapper[4806]: I1204 04:28:58.680283 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dtdft" podUID="c047d3a7-46e0-4f42-ac87-22700753f4f1" containerName="registry-server" containerID="cri-o://f94c31ed94d9c481de0eb00423963c7b57d62b878b592b3d60cbca09d808d41f" gracePeriod=2 Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.369343 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.446653 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckl8w\" (UniqueName: \"kubernetes.io/projected/c047d3a7-46e0-4f42-ac87-22700753f4f1-kube-api-access-ckl8w\") pod \"c047d3a7-46e0-4f42-ac87-22700753f4f1\" (UID: \"c047d3a7-46e0-4f42-ac87-22700753f4f1\") " Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.446747 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c047d3a7-46e0-4f42-ac87-22700753f4f1-catalog-content\") pod \"c047d3a7-46e0-4f42-ac87-22700753f4f1\" (UID: \"c047d3a7-46e0-4f42-ac87-22700753f4f1\") " Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.446779 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c047d3a7-46e0-4f42-ac87-22700753f4f1-utilities\") pod \"c047d3a7-46e0-4f42-ac87-22700753f4f1\" (UID: \"c047d3a7-46e0-4f42-ac87-22700753f4f1\") " Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.447621 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c047d3a7-46e0-4f42-ac87-22700753f4f1-utilities" (OuterVolumeSpecName: "utilities") pod "c047d3a7-46e0-4f42-ac87-22700753f4f1" (UID: "c047d3a7-46e0-4f42-ac87-22700753f4f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.456916 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c047d3a7-46e0-4f42-ac87-22700753f4f1-kube-api-access-ckl8w" (OuterVolumeSpecName: "kube-api-access-ckl8w") pod "c047d3a7-46e0-4f42-ac87-22700753f4f1" (UID: "c047d3a7-46e0-4f42-ac87-22700753f4f1"). InnerVolumeSpecName "kube-api-access-ckl8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.548649 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckl8w\" (UniqueName: \"kubernetes.io/projected/c047d3a7-46e0-4f42-ac87-22700753f4f1-kube-api-access-ckl8w\") on node \"crc\" DevicePath \"\"" Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.548917 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c047d3a7-46e0-4f42-ac87-22700753f4f1-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.576501 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c047d3a7-46e0-4f42-ac87-22700753f4f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c047d3a7-46e0-4f42-ac87-22700753f4f1" (UID: "c047d3a7-46e0-4f42-ac87-22700753f4f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.649906 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c047d3a7-46e0-4f42-ac87-22700753f4f1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.691699 4806 generic.go:334] "Generic (PLEG): container finished" podID="c047d3a7-46e0-4f42-ac87-22700753f4f1" containerID="f94c31ed94d9c481de0eb00423963c7b57d62b878b592b3d60cbca09d808d41f" exitCode=0 Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.692583 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtdft" event={"ID":"c047d3a7-46e0-4f42-ac87-22700753f4f1","Type":"ContainerDied","Data":"f94c31ed94d9c481de0eb00423963c7b57d62b878b592b3d60cbca09d808d41f"} Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.692693 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dtdft" event={"ID":"c047d3a7-46e0-4f42-ac87-22700753f4f1","Type":"ContainerDied","Data":"1b179a058a3aab809f48f6df3bd66379f91d30ce29c3d140aea3f9df42e48214"} Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.692794 4806 scope.go:117] "RemoveContainer" containerID="f94c31ed94d9c481de0eb00423963c7b57d62b878b592b3d60cbca09d808d41f" Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.693087 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dtdft" Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.721432 4806 scope.go:117] "RemoveContainer" containerID="63b98617e04d9a60b350806cc52d29e825df6188083d82c626598a8b0785b837" Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.731426 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dtdft"] Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.741020 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dtdft"] Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.751255 4806 scope.go:117] "RemoveContainer" containerID="21f7e6be351bd95dd978d8e2b08f5538e004755eb96bc2cbe9132a5452b59af4" Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.798092 4806 scope.go:117] "RemoveContainer" containerID="f94c31ed94d9c481de0eb00423963c7b57d62b878b592b3d60cbca09d808d41f" Dec 04 04:28:59 crc kubenswrapper[4806]: E1204 04:28:59.798974 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f94c31ed94d9c481de0eb00423963c7b57d62b878b592b3d60cbca09d808d41f\": container with ID starting with f94c31ed94d9c481de0eb00423963c7b57d62b878b592b3d60cbca09d808d41f not found: ID does not exist" containerID="f94c31ed94d9c481de0eb00423963c7b57d62b878b592b3d60cbca09d808d41f" Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.799015 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f94c31ed94d9c481de0eb00423963c7b57d62b878b592b3d60cbca09d808d41f"} err="failed to get container status \"f94c31ed94d9c481de0eb00423963c7b57d62b878b592b3d60cbca09d808d41f\": rpc error: code = NotFound desc = could not find container \"f94c31ed94d9c481de0eb00423963c7b57d62b878b592b3d60cbca09d808d41f\": container with ID starting with f94c31ed94d9c481de0eb00423963c7b57d62b878b592b3d60cbca09d808d41f not found: ID does not exist" Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.799036 4806 scope.go:117] "RemoveContainer" containerID="63b98617e04d9a60b350806cc52d29e825df6188083d82c626598a8b0785b837" Dec 04 04:28:59 crc kubenswrapper[4806]: E1204 04:28:59.799475 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63b98617e04d9a60b350806cc52d29e825df6188083d82c626598a8b0785b837\": container with ID starting with 63b98617e04d9a60b350806cc52d29e825df6188083d82c626598a8b0785b837 not found: ID does not exist" containerID="63b98617e04d9a60b350806cc52d29e825df6188083d82c626598a8b0785b837" Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.799508 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63b98617e04d9a60b350806cc52d29e825df6188083d82c626598a8b0785b837"} err="failed to get container status \"63b98617e04d9a60b350806cc52d29e825df6188083d82c626598a8b0785b837\": rpc error: code = NotFound desc = could not find container \"63b98617e04d9a60b350806cc52d29e825df6188083d82c626598a8b0785b837\": container with ID starting with 63b98617e04d9a60b350806cc52d29e825df6188083d82c626598a8b0785b837 not found: ID does not exist" Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.799528 4806 scope.go:117] "RemoveContainer" containerID="21f7e6be351bd95dd978d8e2b08f5538e004755eb96bc2cbe9132a5452b59af4" Dec 04 04:28:59 crc kubenswrapper[4806]: E1204 04:28:59.799844 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21f7e6be351bd95dd978d8e2b08f5538e004755eb96bc2cbe9132a5452b59af4\": container with ID starting with 21f7e6be351bd95dd978d8e2b08f5538e004755eb96bc2cbe9132a5452b59af4 not found: ID does not exist" containerID="21f7e6be351bd95dd978d8e2b08f5538e004755eb96bc2cbe9132a5452b59af4" Dec 04 04:28:59 crc kubenswrapper[4806]: I1204 04:28:59.799874 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21f7e6be351bd95dd978d8e2b08f5538e004755eb96bc2cbe9132a5452b59af4"} err="failed to get container status \"21f7e6be351bd95dd978d8e2b08f5538e004755eb96bc2cbe9132a5452b59af4\": rpc error: code = NotFound desc = could not find container \"21f7e6be351bd95dd978d8e2b08f5538e004755eb96bc2cbe9132a5452b59af4\": container with ID starting with 21f7e6be351bd95dd978d8e2b08f5538e004755eb96bc2cbe9132a5452b59af4 not found: ID does not exist" Dec 04 04:29:01 crc kubenswrapper[4806]: I1204 04:29:01.345362 4806 scope.go:117] "RemoveContainer" containerID="bf97c3e4038b06bfe8cd8862506302e6218b0048b8d8a9acd66a8c02e156fedc" Dec 04 04:29:01 crc kubenswrapper[4806]: I1204 04:29:01.444599 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c047d3a7-46e0-4f42-ac87-22700753f4f1" path="/var/lib/kubelet/pods/c047d3a7-46e0-4f42-ac87-22700753f4f1/volumes" Dec 04 04:29:47 crc kubenswrapper[4806]: I1204 04:29:47.116722 4806 generic.go:334] "Generic (PLEG): container finished" podID="cc3788a8-d331-4c5c-a80b-d73e1ef5da9c" containerID="4ab9c98bf2cac8df8c86fbe81f501b5d312346fffe2b4c3e7b31c005d18be3b0" exitCode=0 Dec 04 04:29:47 crc kubenswrapper[4806]: I1204 04:29:47.116868 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" event={"ID":"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c","Type":"ContainerDied","Data":"4ab9c98bf2cac8df8c86fbe81f501b5d312346fffe2b4c3e7b31c005d18be3b0"} Dec 04 04:29:48 crc kubenswrapper[4806]: I1204 04:29:48.578793 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" Dec 04 04:29:48 crc kubenswrapper[4806]: I1204 04:29:48.660778 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf4k5\" (UniqueName: \"kubernetes.io/projected/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-kube-api-access-sf4k5\") pod \"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c\" (UID: \"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c\") " Dec 04 04:29:48 crc kubenswrapper[4806]: I1204 04:29:48.660882 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-ssh-key\") pod \"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c\" (UID: \"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c\") " Dec 04 04:29:48 crc kubenswrapper[4806]: I1204 04:29:48.660947 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-inventory\") pod \"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c\" (UID: \"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c\") " Dec 04 04:29:48 crc kubenswrapper[4806]: I1204 04:29:48.669176 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-kube-api-access-sf4k5" (OuterVolumeSpecName: "kube-api-access-sf4k5") pod "cc3788a8-d331-4c5c-a80b-d73e1ef5da9c" (UID: "cc3788a8-d331-4c5c-a80b-d73e1ef5da9c"). InnerVolumeSpecName "kube-api-access-sf4k5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:29:48 crc kubenswrapper[4806]: I1204 04:29:48.694330 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-inventory" (OuterVolumeSpecName: "inventory") pod "cc3788a8-d331-4c5c-a80b-d73e1ef5da9c" (UID: "cc3788a8-d331-4c5c-a80b-d73e1ef5da9c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:29:48 crc kubenswrapper[4806]: I1204 04:29:48.710173 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cc3788a8-d331-4c5c-a80b-d73e1ef5da9c" (UID: "cc3788a8-d331-4c5c-a80b-d73e1ef5da9c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:29:48 crc kubenswrapper[4806]: I1204 04:29:48.763681 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf4k5\" (UniqueName: \"kubernetes.io/projected/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-kube-api-access-sf4k5\") on node \"crc\" DevicePath \"\"" Dec 04 04:29:48 crc kubenswrapper[4806]: I1204 04:29:48.763727 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:29:48 crc kubenswrapper[4806]: I1204 04:29:48.763740 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc3788a8-d331-4c5c-a80b-d73e1ef5da9c-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.136697 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" event={"ID":"cc3788a8-d331-4c5c-a80b-d73e1ef5da9c","Type":"ContainerDied","Data":"73d892f6d29c122f5e7e16999ff2d9e40362b6e5427120a40ff2d8f9534e71bb"} Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.136745 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73d892f6d29c122f5e7e16999ff2d9e40362b6e5427120a40ff2d8f9534e71bb" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.137109 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-984f8" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.302351 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cddcl"] Dec 04 04:29:49 crc kubenswrapper[4806]: E1204 04:29:49.302840 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc3788a8-d331-4c5c-a80b-d73e1ef5da9c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.302863 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc3788a8-d331-4c5c-a80b-d73e1ef5da9c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 04 04:29:49 crc kubenswrapper[4806]: E1204 04:29:49.302901 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c047d3a7-46e0-4f42-ac87-22700753f4f1" containerName="extract-content" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.302910 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c047d3a7-46e0-4f42-ac87-22700753f4f1" containerName="extract-content" Dec 04 04:29:49 crc kubenswrapper[4806]: E1204 04:29:49.302955 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c047d3a7-46e0-4f42-ac87-22700753f4f1" containerName="extract-utilities" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.302963 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c047d3a7-46e0-4f42-ac87-22700753f4f1" containerName="extract-utilities" Dec 04 04:29:49 crc kubenswrapper[4806]: E1204 04:29:49.302979 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c047d3a7-46e0-4f42-ac87-22700753f4f1" containerName="registry-server" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.302986 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c047d3a7-46e0-4f42-ac87-22700753f4f1" containerName="registry-server" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.303228 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="c047d3a7-46e0-4f42-ac87-22700753f4f1" containerName="registry-server" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.303263 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc3788a8-d331-4c5c-a80b-d73e1ef5da9c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.304114 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.306414 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.307490 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.307562 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-q8mqk" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.307490 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.322123 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cddcl"] Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.375811 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cf2d196e-c288-449d-a106-2ca0b6697582-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cddcl\" (UID: \"cf2d196e-c288-449d-a106-2ca0b6697582\") " pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.375958 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhq2r\" (UniqueName: \"kubernetes.io/projected/cf2d196e-c288-449d-a106-2ca0b6697582-kube-api-access-mhq2r\") pod \"ssh-known-hosts-edpm-deployment-cddcl\" (UID: \"cf2d196e-c288-449d-a106-2ca0b6697582\") " pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.376065 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cf2d196e-c288-449d-a106-2ca0b6697582-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cddcl\" (UID: \"cf2d196e-c288-449d-a106-2ca0b6697582\") " pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.477352 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cf2d196e-c288-449d-a106-2ca0b6697582-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cddcl\" (UID: \"cf2d196e-c288-449d-a106-2ca0b6697582\") " pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.477711 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhq2r\" (UniqueName: \"kubernetes.io/projected/cf2d196e-c288-449d-a106-2ca0b6697582-kube-api-access-mhq2r\") pod \"ssh-known-hosts-edpm-deployment-cddcl\" (UID: \"cf2d196e-c288-449d-a106-2ca0b6697582\") " pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.477819 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cf2d196e-c288-449d-a106-2ca0b6697582-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cddcl\" (UID: \"cf2d196e-c288-449d-a106-2ca0b6697582\") " pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.489502 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cf2d196e-c288-449d-a106-2ca0b6697582-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cddcl\" (UID: \"cf2d196e-c288-449d-a106-2ca0b6697582\") " pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.491800 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cf2d196e-c288-449d-a106-2ca0b6697582-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cddcl\" (UID: \"cf2d196e-c288-449d-a106-2ca0b6697582\") " pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.495771 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhq2r\" (UniqueName: \"kubernetes.io/projected/cf2d196e-c288-449d-a106-2ca0b6697582-kube-api-access-mhq2r\") pod \"ssh-known-hosts-edpm-deployment-cddcl\" (UID: \"cf2d196e-c288-449d-a106-2ca0b6697582\") " pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" Dec 04 04:29:49 crc kubenswrapper[4806]: I1204 04:29:49.622823 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" Dec 04 04:29:50 crc kubenswrapper[4806]: I1204 04:29:50.256559 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cddcl"] Dec 04 04:29:50 crc kubenswrapper[4806]: W1204 04:29:50.300063 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf2d196e_c288_449d_a106_2ca0b6697582.slice/crio-268b08fbadfa96a13891587f8a72c7fa478cf81a6148b618b58731a0b9481f6c WatchSource:0}: Error finding container 268b08fbadfa96a13891587f8a72c7fa478cf81a6148b618b58731a0b9481f6c: Status 404 returned error can't find the container with id 268b08fbadfa96a13891587f8a72c7fa478cf81a6148b618b58731a0b9481f6c Dec 04 04:29:51 crc kubenswrapper[4806]: I1204 04:29:51.157132 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" event={"ID":"cf2d196e-c288-449d-a106-2ca0b6697582","Type":"ContainerStarted","Data":"38b292cc8c8be66282111c97391588a14e9bc1a8efe78a9171fab02d66ca5d34"} Dec 04 04:29:51 crc kubenswrapper[4806]: I1204 04:29:51.157762 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" event={"ID":"cf2d196e-c288-449d-a106-2ca0b6697582","Type":"ContainerStarted","Data":"268b08fbadfa96a13891587f8a72c7fa478cf81a6148b618b58731a0b9481f6c"} Dec 04 04:29:51 crc kubenswrapper[4806]: I1204 04:29:51.184951 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" podStartSLOduration=1.94558022 podStartE2EDuration="2.184906462s" podCreationTimestamp="2025-12-04 04:29:49 +0000 UTC" firstStartedPulling="2025-12-04 04:29:50.314045559 +0000 UTC m=+2105.172558487" lastFinishedPulling="2025-12-04 04:29:50.553371801 +0000 UTC m=+2105.411884729" observedRunningTime="2025-12-04 04:29:51.183036274 +0000 UTC m=+2106.041549202" watchObservedRunningTime="2025-12-04 04:29:51.184906462 +0000 UTC m=+2106.043419400" Dec 04 04:29:57 crc kubenswrapper[4806]: I1204 04:29:57.047672 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:29:57 crc kubenswrapper[4806]: I1204 04:29:57.048558 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:29:59 crc kubenswrapper[4806]: I1204 04:29:59.225327 4806 generic.go:334] "Generic (PLEG): container finished" podID="cf2d196e-c288-449d-a106-2ca0b6697582" containerID="38b292cc8c8be66282111c97391588a14e9bc1a8efe78a9171fab02d66ca5d34" exitCode=0 Dec 04 04:29:59 crc kubenswrapper[4806]: I1204 04:29:59.225407 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" event={"ID":"cf2d196e-c288-449d-a106-2ca0b6697582","Type":"ContainerDied","Data":"38b292cc8c8be66282111c97391588a14e9bc1a8efe78a9171fab02d66ca5d34"} Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.163296 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp"] Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.164826 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp" Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.167621 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.167651 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.235563 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp"] Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.295024 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg2jv\" (UniqueName: \"kubernetes.io/projected/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-kube-api-access-gg2jv\") pod \"collect-profiles-29413710-xt7pp\" (UID: \"e456fba1-e611-4f88-9ead-22a8c3c1bc0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp" Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.295350 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-secret-volume\") pod \"collect-profiles-29413710-xt7pp\" (UID: \"e456fba1-e611-4f88-9ead-22a8c3c1bc0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp" Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.295425 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-config-volume\") pod \"collect-profiles-29413710-xt7pp\" (UID: \"e456fba1-e611-4f88-9ead-22a8c3c1bc0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp" Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.398445 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg2jv\" (UniqueName: \"kubernetes.io/projected/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-kube-api-access-gg2jv\") pod \"collect-profiles-29413710-xt7pp\" (UID: \"e456fba1-e611-4f88-9ead-22a8c3c1bc0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp" Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.398683 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-secret-volume\") pod \"collect-profiles-29413710-xt7pp\" (UID: \"e456fba1-e611-4f88-9ead-22a8c3c1bc0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp" Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.398742 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-config-volume\") pod \"collect-profiles-29413710-xt7pp\" (UID: \"e456fba1-e611-4f88-9ead-22a8c3c1bc0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp" Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.400089 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-config-volume\") pod \"collect-profiles-29413710-xt7pp\" (UID: \"e456fba1-e611-4f88-9ead-22a8c3c1bc0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp" Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.407540 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-secret-volume\") pod \"collect-profiles-29413710-xt7pp\" (UID: \"e456fba1-e611-4f88-9ead-22a8c3c1bc0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp" Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.420817 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg2jv\" (UniqueName: \"kubernetes.io/projected/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-kube-api-access-gg2jv\") pod \"collect-profiles-29413710-xt7pp\" (UID: \"e456fba1-e611-4f88-9ead-22a8c3c1bc0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp" Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.502408 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp" Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.779913 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.913144 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhq2r\" (UniqueName: \"kubernetes.io/projected/cf2d196e-c288-449d-a106-2ca0b6697582-kube-api-access-mhq2r\") pod \"cf2d196e-c288-449d-a106-2ca0b6697582\" (UID: \"cf2d196e-c288-449d-a106-2ca0b6697582\") " Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.913283 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cf2d196e-c288-449d-a106-2ca0b6697582-ssh-key-openstack-edpm-ipam\") pod \"cf2d196e-c288-449d-a106-2ca0b6697582\" (UID: \"cf2d196e-c288-449d-a106-2ca0b6697582\") " Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.913371 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cf2d196e-c288-449d-a106-2ca0b6697582-inventory-0\") pod \"cf2d196e-c288-449d-a106-2ca0b6697582\" (UID: \"cf2d196e-c288-449d-a106-2ca0b6697582\") " Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.921038 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf2d196e-c288-449d-a106-2ca0b6697582-kube-api-access-mhq2r" (OuterVolumeSpecName: "kube-api-access-mhq2r") pod "cf2d196e-c288-449d-a106-2ca0b6697582" (UID: "cf2d196e-c288-449d-a106-2ca0b6697582"). InnerVolumeSpecName "kube-api-access-mhq2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.972218 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf2d196e-c288-449d-a106-2ca0b6697582-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "cf2d196e-c288-449d-a106-2ca0b6697582" (UID: "cf2d196e-c288-449d-a106-2ca0b6697582"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:30:00 crc kubenswrapper[4806]: I1204 04:30:00.974343 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf2d196e-c288-449d-a106-2ca0b6697582-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "cf2d196e-c288-449d-a106-2ca0b6697582" (UID: "cf2d196e-c288-449d-a106-2ca0b6697582"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.016150 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cf2d196e-c288-449d-a106-2ca0b6697582-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.016204 4806 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cf2d196e-c288-449d-a106-2ca0b6697582-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.016215 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhq2r\" (UniqueName: \"kubernetes.io/projected/cf2d196e-c288-449d-a106-2ca0b6697582-kube-api-access-mhq2r\") on node \"crc\" DevicePath \"\"" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.158889 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp"] Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.245809 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp" event={"ID":"e456fba1-e611-4f88-9ead-22a8c3c1bc0b","Type":"ContainerStarted","Data":"c2e2cad45521c8bf3afed009dc9e3ee46208c1a4aac74ac9bd1d546fa2fea281"} Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.248674 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" event={"ID":"cf2d196e-c288-449d-a106-2ca0b6697582","Type":"ContainerDied","Data":"268b08fbadfa96a13891587f8a72c7fa478cf81a6148b618b58731a0b9481f6c"} Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.248738 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cddcl" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.248751 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="268b08fbadfa96a13891587f8a72c7fa478cf81a6148b618b58731a0b9481f6c" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.348204 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667"] Dec 04 04:30:01 crc kubenswrapper[4806]: E1204 04:30:01.348677 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf2d196e-c288-449d-a106-2ca0b6697582" containerName="ssh-known-hosts-edpm-deployment" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.348704 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf2d196e-c288-449d-a106-2ca0b6697582" containerName="ssh-known-hosts-edpm-deployment" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.349015 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf2d196e-c288-449d-a106-2ca0b6697582" containerName="ssh-known-hosts-edpm-deployment" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.349760 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.367078 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667"] Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.371771 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.373018 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.373161 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.373318 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-q8mqk" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.423152 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcd44\" (UniqueName: \"kubernetes.io/projected/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-kube-api-access-tcd44\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25667\" (UID: \"e5f894f3-1fed-46cc-9caa-8f61712e8ab3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.423300 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25667\" (UID: \"e5f894f3-1fed-46cc-9caa-8f61712e8ab3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.423334 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25667\" (UID: \"e5f894f3-1fed-46cc-9caa-8f61712e8ab3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.525163 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcd44\" (UniqueName: \"kubernetes.io/projected/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-kube-api-access-tcd44\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25667\" (UID: \"e5f894f3-1fed-46cc-9caa-8f61712e8ab3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.526690 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25667\" (UID: \"e5f894f3-1fed-46cc-9caa-8f61712e8ab3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.527280 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25667\" (UID: \"e5f894f3-1fed-46cc-9caa-8f61712e8ab3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.532639 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25667\" (UID: \"e5f894f3-1fed-46cc-9caa-8f61712e8ab3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.534747 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25667\" (UID: \"e5f894f3-1fed-46cc-9caa-8f61712e8ab3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.542288 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcd44\" (UniqueName: \"kubernetes.io/projected/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-kube-api-access-tcd44\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-25667\" (UID: \"e5f894f3-1fed-46cc-9caa-8f61712e8ab3\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" Dec 04 04:30:01 crc kubenswrapper[4806]: I1204 04:30:01.674614 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" Dec 04 04:30:02 crc kubenswrapper[4806]: I1204 04:30:02.271862 4806 generic.go:334] "Generic (PLEG): container finished" podID="e456fba1-e611-4f88-9ead-22a8c3c1bc0b" containerID="383c717009bb281750543842f95b30ff9a9ead8910583ddc5ad438fc5e444db1" exitCode=0 Dec 04 04:30:02 crc kubenswrapper[4806]: I1204 04:30:02.271965 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp" event={"ID":"e456fba1-e611-4f88-9ead-22a8c3c1bc0b","Type":"ContainerDied","Data":"383c717009bb281750543842f95b30ff9a9ead8910583ddc5ad438fc5e444db1"} Dec 04 04:30:02 crc kubenswrapper[4806]: I1204 04:30:02.338470 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667"] Dec 04 04:30:03 crc kubenswrapper[4806]: I1204 04:30:03.291854 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" event={"ID":"e5f894f3-1fed-46cc-9caa-8f61712e8ab3","Type":"ContainerStarted","Data":"1d9a32be450129d15114e6251f9ca13e6fbfc2658acee8768d8c0ed8dcb3249f"} Dec 04 04:30:03 crc kubenswrapper[4806]: I1204 04:30:03.292906 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" event={"ID":"e5f894f3-1fed-46cc-9caa-8f61712e8ab3","Type":"ContainerStarted","Data":"d2275a69afe231928fd68a12bdd85ccb53c2de38dacfa9e0de2e8f36a15c6f03"} Dec 04 04:30:03 crc kubenswrapper[4806]: I1204 04:30:03.338322 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" podStartSLOduration=2.129989314 podStartE2EDuration="2.338202752s" podCreationTimestamp="2025-12-04 04:30:01 +0000 UTC" firstStartedPulling="2025-12-04 04:30:02.378455261 +0000 UTC m=+2117.236968189" lastFinishedPulling="2025-12-04 04:30:02.586668679 +0000 UTC m=+2117.445181627" observedRunningTime="2025-12-04 04:30:03.320985284 +0000 UTC m=+2118.179498212" watchObservedRunningTime="2025-12-04 04:30:03.338202752 +0000 UTC m=+2118.196715680" Dec 04 04:30:03 crc kubenswrapper[4806]: I1204 04:30:03.690181 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp" Dec 04 04:30:03 crc kubenswrapper[4806]: I1204 04:30:03.791240 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gg2jv\" (UniqueName: \"kubernetes.io/projected/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-kube-api-access-gg2jv\") pod \"e456fba1-e611-4f88-9ead-22a8c3c1bc0b\" (UID: \"e456fba1-e611-4f88-9ead-22a8c3c1bc0b\") " Dec 04 04:30:03 crc kubenswrapper[4806]: I1204 04:30:03.791670 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-config-volume\") pod \"e456fba1-e611-4f88-9ead-22a8c3c1bc0b\" (UID: \"e456fba1-e611-4f88-9ead-22a8c3c1bc0b\") " Dec 04 04:30:03 crc kubenswrapper[4806]: I1204 04:30:03.791885 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-secret-volume\") pod \"e456fba1-e611-4f88-9ead-22a8c3c1bc0b\" (UID: \"e456fba1-e611-4f88-9ead-22a8c3c1bc0b\") " Dec 04 04:30:03 crc kubenswrapper[4806]: I1204 04:30:03.792452 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-config-volume" (OuterVolumeSpecName: "config-volume") pod "e456fba1-e611-4f88-9ead-22a8c3c1bc0b" (UID: "e456fba1-e611-4f88-9ead-22a8c3c1bc0b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:30:03 crc kubenswrapper[4806]: I1204 04:30:03.798619 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e456fba1-e611-4f88-9ead-22a8c3c1bc0b" (UID: "e456fba1-e611-4f88-9ead-22a8c3c1bc0b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:30:03 crc kubenswrapper[4806]: I1204 04:30:03.801221 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-kube-api-access-gg2jv" (OuterVolumeSpecName: "kube-api-access-gg2jv") pod "e456fba1-e611-4f88-9ead-22a8c3c1bc0b" (UID: "e456fba1-e611-4f88-9ead-22a8c3c1bc0b"). InnerVolumeSpecName "kube-api-access-gg2jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:30:03 crc kubenswrapper[4806]: I1204 04:30:03.894800 4806 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 04:30:03 crc kubenswrapper[4806]: I1204 04:30:03.894860 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gg2jv\" (UniqueName: \"kubernetes.io/projected/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-kube-api-access-gg2jv\") on node \"crc\" DevicePath \"\"" Dec 04 04:30:03 crc kubenswrapper[4806]: I1204 04:30:03.894872 4806 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e456fba1-e611-4f88-9ead-22a8c3c1bc0b-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 04:30:04 crc kubenswrapper[4806]: I1204 04:30:04.300984 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp" event={"ID":"e456fba1-e611-4f88-9ead-22a8c3c1bc0b","Type":"ContainerDied","Data":"c2e2cad45521c8bf3afed009dc9e3ee46208c1a4aac74ac9bd1d546fa2fea281"} Dec 04 04:30:04 crc kubenswrapper[4806]: I1204 04:30:04.301031 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp" Dec 04 04:30:04 crc kubenswrapper[4806]: I1204 04:30:04.301039 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2e2cad45521c8bf3afed009dc9e3ee46208c1a4aac74ac9bd1d546fa2fea281" Dec 04 04:30:04 crc kubenswrapper[4806]: I1204 04:30:04.777132 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k"] Dec 04 04:30:04 crc kubenswrapper[4806]: I1204 04:30:04.785793 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413665-w4p5k"] Dec 04 04:30:05 crc kubenswrapper[4806]: I1204 04:30:05.436309 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4f1c998-fb99-4827-a353-d8b59a309cee" path="/var/lib/kubelet/pods/b4f1c998-fb99-4827-a353-d8b59a309cee/volumes" Dec 04 04:30:12 crc kubenswrapper[4806]: I1204 04:30:12.374764 4806 generic.go:334] "Generic (PLEG): container finished" podID="e5f894f3-1fed-46cc-9caa-8f61712e8ab3" containerID="1d9a32be450129d15114e6251f9ca13e6fbfc2658acee8768d8c0ed8dcb3249f" exitCode=0 Dec 04 04:30:12 crc kubenswrapper[4806]: I1204 04:30:12.375407 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" event={"ID":"e5f894f3-1fed-46cc-9caa-8f61712e8ab3","Type":"ContainerDied","Data":"1d9a32be450129d15114e6251f9ca13e6fbfc2658acee8768d8c0ed8dcb3249f"} Dec 04 04:30:13 crc kubenswrapper[4806]: I1204 04:30:13.854583 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" Dec 04 04:30:13 crc kubenswrapper[4806]: I1204 04:30:13.907661 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcd44\" (UniqueName: \"kubernetes.io/projected/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-kube-api-access-tcd44\") pod \"e5f894f3-1fed-46cc-9caa-8f61712e8ab3\" (UID: \"e5f894f3-1fed-46cc-9caa-8f61712e8ab3\") " Dec 04 04:30:13 crc kubenswrapper[4806]: I1204 04:30:13.907760 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-ssh-key\") pod \"e5f894f3-1fed-46cc-9caa-8f61712e8ab3\" (UID: \"e5f894f3-1fed-46cc-9caa-8f61712e8ab3\") " Dec 04 04:30:13 crc kubenswrapper[4806]: I1204 04:30:13.907841 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-inventory\") pod \"e5f894f3-1fed-46cc-9caa-8f61712e8ab3\" (UID: \"e5f894f3-1fed-46cc-9caa-8f61712e8ab3\") " Dec 04 04:30:13 crc kubenswrapper[4806]: I1204 04:30:13.916424 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-kube-api-access-tcd44" (OuterVolumeSpecName: "kube-api-access-tcd44") pod "e5f894f3-1fed-46cc-9caa-8f61712e8ab3" (UID: "e5f894f3-1fed-46cc-9caa-8f61712e8ab3"). InnerVolumeSpecName "kube-api-access-tcd44". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:30:13 crc kubenswrapper[4806]: I1204 04:30:13.944408 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-inventory" (OuterVolumeSpecName: "inventory") pod "e5f894f3-1fed-46cc-9caa-8f61712e8ab3" (UID: "e5f894f3-1fed-46cc-9caa-8f61712e8ab3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:30:13 crc kubenswrapper[4806]: I1204 04:30:13.950545 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e5f894f3-1fed-46cc-9caa-8f61712e8ab3" (UID: "e5f894f3-1fed-46cc-9caa-8f61712e8ab3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.011556 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcd44\" (UniqueName: \"kubernetes.io/projected/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-kube-api-access-tcd44\") on node \"crc\" DevicePath \"\"" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.011600 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.011609 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e5f894f3-1fed-46cc-9caa-8f61712e8ab3-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.395392 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" event={"ID":"e5f894f3-1fed-46cc-9caa-8f61712e8ab3","Type":"ContainerDied","Data":"d2275a69afe231928fd68a12bdd85ccb53c2de38dacfa9e0de2e8f36a15c6f03"} Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.396001 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2275a69afe231928fd68a12bdd85ccb53c2de38dacfa9e0de2e8f36a15c6f03" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.395439 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-25667" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.507736 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln"] Dec 04 04:30:14 crc kubenswrapper[4806]: E1204 04:30:14.508338 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e456fba1-e611-4f88-9ead-22a8c3c1bc0b" containerName="collect-profiles" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.508359 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e456fba1-e611-4f88-9ead-22a8c3c1bc0b" containerName="collect-profiles" Dec 04 04:30:14 crc kubenswrapper[4806]: E1204 04:30:14.508387 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f894f3-1fed-46cc-9caa-8f61712e8ab3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.508396 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f894f3-1fed-46cc-9caa-8f61712e8ab3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.508653 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e456fba1-e611-4f88-9ead-22a8c3c1bc0b" containerName="collect-profiles" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.508679 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5f894f3-1fed-46cc-9caa-8f61712e8ab3" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.509396 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.512909 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-q8mqk" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.513121 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.513286 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.517377 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.524080 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln"] Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.620303 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fprln\" (UID: \"a74530eb-f7f4-40f7-b1fc-c66cba904ce5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.620399 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrjfj\" (UniqueName: \"kubernetes.io/projected/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-kube-api-access-rrjfj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fprln\" (UID: \"a74530eb-f7f4-40f7-b1fc-c66cba904ce5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.620425 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fprln\" (UID: \"a74530eb-f7f4-40f7-b1fc-c66cba904ce5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.722054 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fprln\" (UID: \"a74530eb-f7f4-40f7-b1fc-c66cba904ce5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.722526 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrjfj\" (UniqueName: \"kubernetes.io/projected/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-kube-api-access-rrjfj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fprln\" (UID: \"a74530eb-f7f4-40f7-b1fc-c66cba904ce5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.722694 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fprln\" (UID: \"a74530eb-f7f4-40f7-b1fc-c66cba904ce5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.727605 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fprln\" (UID: \"a74530eb-f7f4-40f7-b1fc-c66cba904ce5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.727605 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fprln\" (UID: \"a74530eb-f7f4-40f7-b1fc-c66cba904ce5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.744625 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrjfj\" (UniqueName: \"kubernetes.io/projected/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-kube-api-access-rrjfj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fprln\" (UID: \"a74530eb-f7f4-40f7-b1fc-c66cba904ce5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" Dec 04 04:30:14 crc kubenswrapper[4806]: I1204 04:30:14.828982 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" Dec 04 04:30:15 crc kubenswrapper[4806]: I1204 04:30:15.374400 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln"] Dec 04 04:30:15 crc kubenswrapper[4806]: I1204 04:30:15.412244 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" event={"ID":"a74530eb-f7f4-40f7-b1fc-c66cba904ce5","Type":"ContainerStarted","Data":"1a95f395ccfc9b9e1a26bb8e09f5d5225d074da12549513e506aa51343eb4126"} Dec 04 04:30:16 crc kubenswrapper[4806]: I1204 04:30:16.421121 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" event={"ID":"a74530eb-f7f4-40f7-b1fc-c66cba904ce5","Type":"ContainerStarted","Data":"5ec80d0f1c8ad274ca5b747e6723ac3d2808a34fbd6e1bc8706192402d4fcfaf"} Dec 04 04:30:16 crc kubenswrapper[4806]: I1204 04:30:16.449138 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" podStartSLOduration=2.281130586 podStartE2EDuration="2.449109527s" podCreationTimestamp="2025-12-04 04:30:14 +0000 UTC" firstStartedPulling="2025-12-04 04:30:15.393174308 +0000 UTC m=+2130.251687236" lastFinishedPulling="2025-12-04 04:30:15.561153249 +0000 UTC m=+2130.419666177" observedRunningTime="2025-12-04 04:30:16.441396425 +0000 UTC m=+2131.299909343" watchObservedRunningTime="2025-12-04 04:30:16.449109527 +0000 UTC m=+2131.307622455" Dec 04 04:30:26 crc kubenswrapper[4806]: I1204 04:30:26.524755 4806 generic.go:334] "Generic (PLEG): container finished" podID="a74530eb-f7f4-40f7-b1fc-c66cba904ce5" containerID="5ec80d0f1c8ad274ca5b747e6723ac3d2808a34fbd6e1bc8706192402d4fcfaf" exitCode=0 Dec 04 04:30:26 crc kubenswrapper[4806]: I1204 04:30:26.524849 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" event={"ID":"a74530eb-f7f4-40f7-b1fc-c66cba904ce5","Type":"ContainerDied","Data":"5ec80d0f1c8ad274ca5b747e6723ac3d2808a34fbd6e1bc8706192402d4fcfaf"} Dec 04 04:30:27 crc kubenswrapper[4806]: I1204 04:30:27.047622 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:30:27 crc kubenswrapper[4806]: I1204 04:30:27.047685 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:30:27 crc kubenswrapper[4806]: I1204 04:30:27.967495 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.104136 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-inventory\") pod \"a74530eb-f7f4-40f7-b1fc-c66cba904ce5\" (UID: \"a74530eb-f7f4-40f7-b1fc-c66cba904ce5\") " Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.104197 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-ssh-key\") pod \"a74530eb-f7f4-40f7-b1fc-c66cba904ce5\" (UID: \"a74530eb-f7f4-40f7-b1fc-c66cba904ce5\") " Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.104233 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrjfj\" (UniqueName: \"kubernetes.io/projected/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-kube-api-access-rrjfj\") pod \"a74530eb-f7f4-40f7-b1fc-c66cba904ce5\" (UID: \"a74530eb-f7f4-40f7-b1fc-c66cba904ce5\") " Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.121375 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-kube-api-access-rrjfj" (OuterVolumeSpecName: "kube-api-access-rrjfj") pod "a74530eb-f7f4-40f7-b1fc-c66cba904ce5" (UID: "a74530eb-f7f4-40f7-b1fc-c66cba904ce5"). InnerVolumeSpecName "kube-api-access-rrjfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.135253 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-inventory" (OuterVolumeSpecName: "inventory") pod "a74530eb-f7f4-40f7-b1fc-c66cba904ce5" (UID: "a74530eb-f7f4-40f7-b1fc-c66cba904ce5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.140445 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a74530eb-f7f4-40f7-b1fc-c66cba904ce5" (UID: "a74530eb-f7f4-40f7-b1fc-c66cba904ce5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.206512 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.206553 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.206566 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrjfj\" (UniqueName: \"kubernetes.io/projected/a74530eb-f7f4-40f7-b1fc-c66cba904ce5-kube-api-access-rrjfj\") on node \"crc\" DevicePath \"\"" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.542946 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" event={"ID":"a74530eb-f7f4-40f7-b1fc-c66cba904ce5","Type":"ContainerDied","Data":"1a95f395ccfc9b9e1a26bb8e09f5d5225d074da12549513e506aa51343eb4126"} Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.542988 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a95f395ccfc9b9e1a26bb8e09f5d5225d074da12549513e506aa51343eb4126" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.543244 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fprln" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.665638 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j"] Dec 04 04:30:28 crc kubenswrapper[4806]: E1204 04:30:28.672188 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a74530eb-f7f4-40f7-b1fc-c66cba904ce5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.672326 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a74530eb-f7f4-40f7-b1fc-c66cba904ce5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.672676 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="a74530eb-f7f4-40f7-b1fc-c66cba904ce5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.673559 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.676309 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.676551 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.678123 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-q8mqk" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.678349 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.679425 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.682993 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.683126 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.683266 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.689778 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j"] Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.817784 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.818195 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.818353 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.818548 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.818702 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.818821 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.818983 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.819145 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.819285 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.819423 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.819554 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r29w6\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-kube-api-access-r29w6\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.819692 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.819845 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.819986 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.921307 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.921365 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.921392 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.921422 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.921458 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.921494 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.921524 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r29w6\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-kube-api-access-r29w6\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.921563 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.921582 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.921598 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.921627 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.921647 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.921666 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.921729 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.926634 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.929876 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.933610 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.933988 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.935301 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.937157 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.937273 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.937231 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.937528 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.938178 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.939218 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r29w6\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-kube-api-access-r29w6\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.943594 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.946509 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:28 crc kubenswrapper[4806]: I1204 04:30:28.948309 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:29 crc kubenswrapper[4806]: I1204 04:30:29.005628 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:30:29 crc kubenswrapper[4806]: I1204 04:30:29.656140 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j"] Dec 04 04:30:30 crc kubenswrapper[4806]: I1204 04:30:30.560257 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" event={"ID":"f3284126-2542-43fc-8e96-829f03db14bc","Type":"ContainerStarted","Data":"6427e5f789e420cea02a6c50b7112e6c6d317e59a4227df134ae4d6709ea4b08"} Dec 04 04:30:30 crc kubenswrapper[4806]: I1204 04:30:30.560673 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" event={"ID":"f3284126-2542-43fc-8e96-829f03db14bc","Type":"ContainerStarted","Data":"2fe7285d98b4700503126fbe1d9a95f8e10b8555b76547f93ef11f6905400017"} Dec 04 04:30:30 crc kubenswrapper[4806]: I1204 04:30:30.588510 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" podStartSLOduration=2.371342852 podStartE2EDuration="2.588491537s" podCreationTimestamp="2025-12-04 04:30:28 +0000 UTC" firstStartedPulling="2025-12-04 04:30:29.669908369 +0000 UTC m=+2144.528421297" lastFinishedPulling="2025-12-04 04:30:29.887057044 +0000 UTC m=+2144.745569982" observedRunningTime="2025-12-04 04:30:30.583329545 +0000 UTC m=+2145.441842473" watchObservedRunningTime="2025-12-04 04:30:30.588491537 +0000 UTC m=+2145.447004465" Dec 04 04:30:57 crc kubenswrapper[4806]: I1204 04:30:57.047546 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:30:57 crc kubenswrapper[4806]: I1204 04:30:57.048126 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:30:57 crc kubenswrapper[4806]: I1204 04:30:57.048175 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 04:30:57 crc kubenswrapper[4806]: I1204 04:30:57.048876 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 04:30:57 crc kubenswrapper[4806]: I1204 04:30:57.048945 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" gracePeriod=600 Dec 04 04:30:57 crc kubenswrapper[4806]: E1204 04:30:57.175211 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:30:57 crc kubenswrapper[4806]: I1204 04:30:57.794806 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" exitCode=0 Dec 04 04:30:57 crc kubenswrapper[4806]: I1204 04:30:57.794852 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513"} Dec 04 04:30:57 crc kubenswrapper[4806]: I1204 04:30:57.794907 4806 scope.go:117] "RemoveContainer" containerID="c15d4fe56702198504fcf9c93df44bc149f0a2bbdf70116fa6a918971b65cecc" Dec 04 04:30:57 crc kubenswrapper[4806]: I1204 04:30:57.797265 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:30:57 crc kubenswrapper[4806]: E1204 04:30:57.798032 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:31:01 crc kubenswrapper[4806]: I1204 04:31:01.493324 4806 scope.go:117] "RemoveContainer" containerID="fa31bd1e6d4845d737a13097baa8f599989c9956742e0a3a2bc54315edef6652" Dec 04 04:31:03 crc kubenswrapper[4806]: I1204 04:31:03.914537 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hx482"] Dec 04 04:31:03 crc kubenswrapper[4806]: I1204 04:31:03.918001 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:03 crc kubenswrapper[4806]: I1204 04:31:03.937291 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hx482"] Dec 04 04:31:04 crc kubenswrapper[4806]: I1204 04:31:04.022939 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4wrx\" (UniqueName: \"kubernetes.io/projected/fb2d3782-54c2-428b-b056-1e5ebd82323e-kube-api-access-n4wrx\") pod \"certified-operators-hx482\" (UID: \"fb2d3782-54c2-428b-b056-1e5ebd82323e\") " pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:04 crc kubenswrapper[4806]: I1204 04:31:04.023032 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb2d3782-54c2-428b-b056-1e5ebd82323e-utilities\") pod \"certified-operators-hx482\" (UID: \"fb2d3782-54c2-428b-b056-1e5ebd82323e\") " pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:04 crc kubenswrapper[4806]: I1204 04:31:04.023074 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb2d3782-54c2-428b-b056-1e5ebd82323e-catalog-content\") pod \"certified-operators-hx482\" (UID: \"fb2d3782-54c2-428b-b056-1e5ebd82323e\") " pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:04 crc kubenswrapper[4806]: I1204 04:31:04.125394 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4wrx\" (UniqueName: \"kubernetes.io/projected/fb2d3782-54c2-428b-b056-1e5ebd82323e-kube-api-access-n4wrx\") pod \"certified-operators-hx482\" (UID: \"fb2d3782-54c2-428b-b056-1e5ebd82323e\") " pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:04 crc kubenswrapper[4806]: I1204 04:31:04.125500 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb2d3782-54c2-428b-b056-1e5ebd82323e-utilities\") pod \"certified-operators-hx482\" (UID: \"fb2d3782-54c2-428b-b056-1e5ebd82323e\") " pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:04 crc kubenswrapper[4806]: I1204 04:31:04.125565 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb2d3782-54c2-428b-b056-1e5ebd82323e-catalog-content\") pod \"certified-operators-hx482\" (UID: \"fb2d3782-54c2-428b-b056-1e5ebd82323e\") " pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:04 crc kubenswrapper[4806]: I1204 04:31:04.126241 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb2d3782-54c2-428b-b056-1e5ebd82323e-catalog-content\") pod \"certified-operators-hx482\" (UID: \"fb2d3782-54c2-428b-b056-1e5ebd82323e\") " pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:04 crc kubenswrapper[4806]: I1204 04:31:04.126991 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb2d3782-54c2-428b-b056-1e5ebd82323e-utilities\") pod \"certified-operators-hx482\" (UID: \"fb2d3782-54c2-428b-b056-1e5ebd82323e\") " pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:04 crc kubenswrapper[4806]: I1204 04:31:04.158504 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4wrx\" (UniqueName: \"kubernetes.io/projected/fb2d3782-54c2-428b-b056-1e5ebd82323e-kube-api-access-n4wrx\") pod \"certified-operators-hx482\" (UID: \"fb2d3782-54c2-428b-b056-1e5ebd82323e\") " pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:04 crc kubenswrapper[4806]: I1204 04:31:04.243309 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:04 crc kubenswrapper[4806]: I1204 04:31:04.927461 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hx482"] Dec 04 04:31:05 crc kubenswrapper[4806]: I1204 04:31:05.880118 4806 generic.go:334] "Generic (PLEG): container finished" podID="fb2d3782-54c2-428b-b056-1e5ebd82323e" containerID="199e5efa4c1de85529465851826c11209370923f17219a5f190e7acc806b4c73" exitCode=0 Dec 04 04:31:05 crc kubenswrapper[4806]: I1204 04:31:05.880228 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hx482" event={"ID":"fb2d3782-54c2-428b-b056-1e5ebd82323e","Type":"ContainerDied","Data":"199e5efa4c1de85529465851826c11209370923f17219a5f190e7acc806b4c73"} Dec 04 04:31:05 crc kubenswrapper[4806]: I1204 04:31:05.881042 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hx482" event={"ID":"fb2d3782-54c2-428b-b056-1e5ebd82323e","Type":"ContainerStarted","Data":"089d73c2b994c734ecaf1ed786a485a7d62a307fe63990894b649a4a2473cd09"} Dec 04 04:31:06 crc kubenswrapper[4806]: I1204 04:31:06.896112 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hx482" event={"ID":"fb2d3782-54c2-428b-b056-1e5ebd82323e","Type":"ContainerStarted","Data":"29f8e9daa39b34207c4900b5ed5089401b438a4bd91b12ce664dfb8f5c46abd7"} Dec 04 04:31:07 crc kubenswrapper[4806]: I1204 04:31:07.909833 4806 generic.go:334] "Generic (PLEG): container finished" podID="fb2d3782-54c2-428b-b056-1e5ebd82323e" containerID="29f8e9daa39b34207c4900b5ed5089401b438a4bd91b12ce664dfb8f5c46abd7" exitCode=0 Dec 04 04:31:07 crc kubenswrapper[4806]: I1204 04:31:07.910048 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hx482" event={"ID":"fb2d3782-54c2-428b-b056-1e5ebd82323e","Type":"ContainerDied","Data":"29f8e9daa39b34207c4900b5ed5089401b438a4bd91b12ce664dfb8f5c46abd7"} Dec 04 04:31:09 crc kubenswrapper[4806]: I1204 04:31:09.931175 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hx482" event={"ID":"fb2d3782-54c2-428b-b056-1e5ebd82323e","Type":"ContainerStarted","Data":"d2b3012fcbf809c142237e1c91510ded21e20feb5a811a231266b554d26fb82c"} Dec 04 04:31:09 crc kubenswrapper[4806]: I1204 04:31:09.967355 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hx482" podStartSLOduration=3.9454863749999998 podStartE2EDuration="6.96732304s" podCreationTimestamp="2025-12-04 04:31:03 +0000 UTC" firstStartedPulling="2025-12-04 04:31:05.883913879 +0000 UTC m=+2180.742426817" lastFinishedPulling="2025-12-04 04:31:08.905750554 +0000 UTC m=+2183.764263482" observedRunningTime="2025-12-04 04:31:09.961465996 +0000 UTC m=+2184.819978924" watchObservedRunningTime="2025-12-04 04:31:09.96732304 +0000 UTC m=+2184.825835968" Dec 04 04:31:10 crc kubenswrapper[4806]: I1204 04:31:10.425459 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:31:10 crc kubenswrapper[4806]: E1204 04:31:10.425685 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:31:14 crc kubenswrapper[4806]: I1204 04:31:14.244418 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:14 crc kubenswrapper[4806]: I1204 04:31:14.244884 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:14 crc kubenswrapper[4806]: I1204 04:31:14.303673 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:14 crc kubenswrapper[4806]: I1204 04:31:14.976108 4806 generic.go:334] "Generic (PLEG): container finished" podID="f3284126-2542-43fc-8e96-829f03db14bc" containerID="6427e5f789e420cea02a6c50b7112e6c6d317e59a4227df134ae4d6709ea4b08" exitCode=0 Dec 04 04:31:14 crc kubenswrapper[4806]: I1204 04:31:14.976156 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" event={"ID":"f3284126-2542-43fc-8e96-829f03db14bc","Type":"ContainerDied","Data":"6427e5f789e420cea02a6c50b7112e6c6d317e59a4227df134ae4d6709ea4b08"} Dec 04 04:31:15 crc kubenswrapper[4806]: I1204 04:31:15.033211 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:15 crc kubenswrapper[4806]: I1204 04:31:15.086973 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hx482"] Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.547308 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.709490 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-neutron-metadata-combined-ca-bundle\") pod \"f3284126-2542-43fc-8e96-829f03db14bc\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.709533 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-telemetry-combined-ca-bundle\") pod \"f3284126-2542-43fc-8e96-829f03db14bc\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.709562 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"f3284126-2542-43fc-8e96-829f03db14bc\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.709595 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r29w6\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-kube-api-access-r29w6\") pod \"f3284126-2542-43fc-8e96-829f03db14bc\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.709650 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-nova-combined-ca-bundle\") pod \"f3284126-2542-43fc-8e96-829f03db14bc\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.709698 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"f3284126-2542-43fc-8e96-829f03db14bc\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.709729 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-libvirt-combined-ca-bundle\") pod \"f3284126-2542-43fc-8e96-829f03db14bc\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.709809 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-ssh-key\") pod \"f3284126-2542-43fc-8e96-829f03db14bc\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.709826 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-ovn-default-certs-0\") pod \"f3284126-2542-43fc-8e96-829f03db14bc\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.709878 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-repo-setup-combined-ca-bundle\") pod \"f3284126-2542-43fc-8e96-829f03db14bc\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.709902 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-bootstrap-combined-ca-bundle\") pod \"f3284126-2542-43fc-8e96-829f03db14bc\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.709946 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-ovn-combined-ca-bundle\") pod \"f3284126-2542-43fc-8e96-829f03db14bc\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.709997 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"f3284126-2542-43fc-8e96-829f03db14bc\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.710053 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-inventory\") pod \"f3284126-2542-43fc-8e96-829f03db14bc\" (UID: \"f3284126-2542-43fc-8e96-829f03db14bc\") " Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.719701 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-kube-api-access-r29w6" (OuterVolumeSpecName: "kube-api-access-r29w6") pod "f3284126-2542-43fc-8e96-829f03db14bc" (UID: "f3284126-2542-43fc-8e96-829f03db14bc"). InnerVolumeSpecName "kube-api-access-r29w6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.737168 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "f3284126-2542-43fc-8e96-829f03db14bc" (UID: "f3284126-2542-43fc-8e96-829f03db14bc"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.739184 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "f3284126-2542-43fc-8e96-829f03db14bc" (UID: "f3284126-2542-43fc-8e96-829f03db14bc"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.742183 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "f3284126-2542-43fc-8e96-829f03db14bc" (UID: "f3284126-2542-43fc-8e96-829f03db14bc"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.742353 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "f3284126-2542-43fc-8e96-829f03db14bc" (UID: "f3284126-2542-43fc-8e96-829f03db14bc"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.742409 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "f3284126-2542-43fc-8e96-829f03db14bc" (UID: "f3284126-2542-43fc-8e96-829f03db14bc"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.742459 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "f3284126-2542-43fc-8e96-829f03db14bc" (UID: "f3284126-2542-43fc-8e96-829f03db14bc"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.742508 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "f3284126-2542-43fc-8e96-829f03db14bc" (UID: "f3284126-2542-43fc-8e96-829f03db14bc"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.742597 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "f3284126-2542-43fc-8e96-829f03db14bc" (UID: "f3284126-2542-43fc-8e96-829f03db14bc"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.742652 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "f3284126-2542-43fc-8e96-829f03db14bc" (UID: "f3284126-2542-43fc-8e96-829f03db14bc"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.755158 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "f3284126-2542-43fc-8e96-829f03db14bc" (UID: "f3284126-2542-43fc-8e96-829f03db14bc"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.783444 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "f3284126-2542-43fc-8e96-829f03db14bc" (UID: "f3284126-2542-43fc-8e96-829f03db14bc"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.790820 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-inventory" (OuterVolumeSpecName: "inventory") pod "f3284126-2542-43fc-8e96-829f03db14bc" (UID: "f3284126-2542-43fc-8e96-829f03db14bc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.805052 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f3284126-2542-43fc-8e96-829f03db14bc" (UID: "f3284126-2542-43fc-8e96-829f03db14bc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.811961 4806 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.812000 4806 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.812014 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.812031 4806 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.812044 4806 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.812059 4806 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.812073 4806 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.812087 4806 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.812098 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.812111 4806 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.812126 4806 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.812142 4806 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.812155 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r29w6\" (UniqueName: \"kubernetes.io/projected/f3284126-2542-43fc-8e96-829f03db14bc-kube-api-access-r29w6\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:16 crc kubenswrapper[4806]: I1204 04:31:16.812167 4806 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3284126-2542-43fc-8e96-829f03db14bc-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.001528 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.001536 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j" event={"ID":"f3284126-2542-43fc-8e96-829f03db14bc","Type":"ContainerDied","Data":"2fe7285d98b4700503126fbe1d9a95f8e10b8555b76547f93ef11f6905400017"} Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.001587 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fe7285d98b4700503126fbe1d9a95f8e10b8555b76547f93ef11f6905400017" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.001640 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hx482" podUID="fb2d3782-54c2-428b-b056-1e5ebd82323e" containerName="registry-server" containerID="cri-o://d2b3012fcbf809c142237e1c91510ded21e20feb5a811a231266b554d26fb82c" gracePeriod=2 Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.138521 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2"] Dec 04 04:31:17 crc kubenswrapper[4806]: E1204 04:31:17.138955 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3284126-2542-43fc-8e96-829f03db14bc" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.138981 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3284126-2542-43fc-8e96-829f03db14bc" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.139191 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3284126-2542-43fc-8e96-829f03db14bc" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.140123 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.143300 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-q8mqk" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.143330 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.143347 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.143612 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.143743 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.192817 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2"] Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.226302 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b4d109a9-916a-41b6-9008-15291d44a0a4-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mbhj2\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.226366 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mbhj2\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.226423 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mbhj2\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.226464 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s94l\" (UniqueName: \"kubernetes.io/projected/b4d109a9-916a-41b6-9008-15291d44a0a4-kube-api-access-5s94l\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mbhj2\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.226555 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mbhj2\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.329245 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mbhj2\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.329808 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s94l\" (UniqueName: \"kubernetes.io/projected/b4d109a9-916a-41b6-9008-15291d44a0a4-kube-api-access-5s94l\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mbhj2\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.329942 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mbhj2\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.330007 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b4d109a9-916a-41b6-9008-15291d44a0a4-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mbhj2\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.330033 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mbhj2\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.331427 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b4d109a9-916a-41b6-9008-15291d44a0a4-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mbhj2\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.337137 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mbhj2\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.338737 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mbhj2\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.339199 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mbhj2\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.349965 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s94l\" (UniqueName: \"kubernetes.io/projected/b4d109a9-916a-41b6-9008-15291d44a0a4-kube-api-access-5s94l\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-mbhj2\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.506713 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.518218 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.636441 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb2d3782-54c2-428b-b056-1e5ebd82323e-catalog-content\") pod \"fb2d3782-54c2-428b-b056-1e5ebd82323e\" (UID: \"fb2d3782-54c2-428b-b056-1e5ebd82323e\") " Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.636602 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4wrx\" (UniqueName: \"kubernetes.io/projected/fb2d3782-54c2-428b-b056-1e5ebd82323e-kube-api-access-n4wrx\") pod \"fb2d3782-54c2-428b-b056-1e5ebd82323e\" (UID: \"fb2d3782-54c2-428b-b056-1e5ebd82323e\") " Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.636953 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb2d3782-54c2-428b-b056-1e5ebd82323e-utilities\") pod \"fb2d3782-54c2-428b-b056-1e5ebd82323e\" (UID: \"fb2d3782-54c2-428b-b056-1e5ebd82323e\") " Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.643105 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb2d3782-54c2-428b-b056-1e5ebd82323e-utilities" (OuterVolumeSpecName: "utilities") pod "fb2d3782-54c2-428b-b056-1e5ebd82323e" (UID: "fb2d3782-54c2-428b-b056-1e5ebd82323e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.647260 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb2d3782-54c2-428b-b056-1e5ebd82323e-kube-api-access-n4wrx" (OuterVolumeSpecName: "kube-api-access-n4wrx") pod "fb2d3782-54c2-428b-b056-1e5ebd82323e" (UID: "fb2d3782-54c2-428b-b056-1e5ebd82323e"). InnerVolumeSpecName "kube-api-access-n4wrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.710745 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb2d3782-54c2-428b-b056-1e5ebd82323e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fb2d3782-54c2-428b-b056-1e5ebd82323e" (UID: "fb2d3782-54c2-428b-b056-1e5ebd82323e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.740546 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb2d3782-54c2-428b-b056-1e5ebd82323e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.740607 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4wrx\" (UniqueName: \"kubernetes.io/projected/fb2d3782-54c2-428b-b056-1e5ebd82323e-kube-api-access-n4wrx\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:17 crc kubenswrapper[4806]: I1204 04:31:17.740626 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb2d3782-54c2-428b-b056-1e5ebd82323e-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:31:18 crc kubenswrapper[4806]: I1204 04:31:18.019297 4806 generic.go:334] "Generic (PLEG): container finished" podID="fb2d3782-54c2-428b-b056-1e5ebd82323e" containerID="d2b3012fcbf809c142237e1c91510ded21e20feb5a811a231266b554d26fb82c" exitCode=0 Dec 04 04:31:18 crc kubenswrapper[4806]: I1204 04:31:18.019352 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hx482" Dec 04 04:31:18 crc kubenswrapper[4806]: I1204 04:31:18.019345 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hx482" event={"ID":"fb2d3782-54c2-428b-b056-1e5ebd82323e","Type":"ContainerDied","Data":"d2b3012fcbf809c142237e1c91510ded21e20feb5a811a231266b554d26fb82c"} Dec 04 04:31:18 crc kubenswrapper[4806]: I1204 04:31:18.019450 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hx482" event={"ID":"fb2d3782-54c2-428b-b056-1e5ebd82323e","Type":"ContainerDied","Data":"089d73c2b994c734ecaf1ed786a485a7d62a307fe63990894b649a4a2473cd09"} Dec 04 04:31:18 crc kubenswrapper[4806]: I1204 04:31:18.019492 4806 scope.go:117] "RemoveContainer" containerID="d2b3012fcbf809c142237e1c91510ded21e20feb5a811a231266b554d26fb82c" Dec 04 04:31:18 crc kubenswrapper[4806]: I1204 04:31:18.072275 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hx482"] Dec 04 04:31:18 crc kubenswrapper[4806]: I1204 04:31:18.077104 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hx482"] Dec 04 04:31:18 crc kubenswrapper[4806]: I1204 04:31:18.085316 4806 scope.go:117] "RemoveContainer" containerID="29f8e9daa39b34207c4900b5ed5089401b438a4bd91b12ce664dfb8f5c46abd7" Dec 04 04:31:18 crc kubenswrapper[4806]: I1204 04:31:18.120999 4806 scope.go:117] "RemoveContainer" containerID="199e5efa4c1de85529465851826c11209370923f17219a5f190e7acc806b4c73" Dec 04 04:31:18 crc kubenswrapper[4806]: I1204 04:31:18.134693 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2"] Dec 04 04:31:18 crc kubenswrapper[4806]: W1204 04:31:18.145886 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4d109a9_916a_41b6_9008_15291d44a0a4.slice/crio-7ead933d0b0851ad84191dae2f4c500d57ce85e00fbecb2734bf8b58b5a20c3b WatchSource:0}: Error finding container 7ead933d0b0851ad84191dae2f4c500d57ce85e00fbecb2734bf8b58b5a20c3b: Status 404 returned error can't find the container with id 7ead933d0b0851ad84191dae2f4c500d57ce85e00fbecb2734bf8b58b5a20c3b Dec 04 04:31:18 crc kubenswrapper[4806]: I1204 04:31:18.163087 4806 scope.go:117] "RemoveContainer" containerID="d2b3012fcbf809c142237e1c91510ded21e20feb5a811a231266b554d26fb82c" Dec 04 04:31:18 crc kubenswrapper[4806]: E1204 04:31:18.166673 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2b3012fcbf809c142237e1c91510ded21e20feb5a811a231266b554d26fb82c\": container with ID starting with d2b3012fcbf809c142237e1c91510ded21e20feb5a811a231266b554d26fb82c not found: ID does not exist" containerID="d2b3012fcbf809c142237e1c91510ded21e20feb5a811a231266b554d26fb82c" Dec 04 04:31:18 crc kubenswrapper[4806]: I1204 04:31:18.166731 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2b3012fcbf809c142237e1c91510ded21e20feb5a811a231266b554d26fb82c"} err="failed to get container status \"d2b3012fcbf809c142237e1c91510ded21e20feb5a811a231266b554d26fb82c\": rpc error: code = NotFound desc = could not find container \"d2b3012fcbf809c142237e1c91510ded21e20feb5a811a231266b554d26fb82c\": container with ID starting with d2b3012fcbf809c142237e1c91510ded21e20feb5a811a231266b554d26fb82c not found: ID does not exist" Dec 04 04:31:18 crc kubenswrapper[4806]: I1204 04:31:18.166767 4806 scope.go:117] "RemoveContainer" containerID="29f8e9daa39b34207c4900b5ed5089401b438a4bd91b12ce664dfb8f5c46abd7" Dec 04 04:31:18 crc kubenswrapper[4806]: E1204 04:31:18.167426 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29f8e9daa39b34207c4900b5ed5089401b438a4bd91b12ce664dfb8f5c46abd7\": container with ID starting with 29f8e9daa39b34207c4900b5ed5089401b438a4bd91b12ce664dfb8f5c46abd7 not found: ID does not exist" containerID="29f8e9daa39b34207c4900b5ed5089401b438a4bd91b12ce664dfb8f5c46abd7" Dec 04 04:31:18 crc kubenswrapper[4806]: I1204 04:31:18.167536 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29f8e9daa39b34207c4900b5ed5089401b438a4bd91b12ce664dfb8f5c46abd7"} err="failed to get container status \"29f8e9daa39b34207c4900b5ed5089401b438a4bd91b12ce664dfb8f5c46abd7\": rpc error: code = NotFound desc = could not find container \"29f8e9daa39b34207c4900b5ed5089401b438a4bd91b12ce664dfb8f5c46abd7\": container with ID starting with 29f8e9daa39b34207c4900b5ed5089401b438a4bd91b12ce664dfb8f5c46abd7 not found: ID does not exist" Dec 04 04:31:18 crc kubenswrapper[4806]: I1204 04:31:18.167582 4806 scope.go:117] "RemoveContainer" containerID="199e5efa4c1de85529465851826c11209370923f17219a5f190e7acc806b4c73" Dec 04 04:31:18 crc kubenswrapper[4806]: E1204 04:31:18.169034 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"199e5efa4c1de85529465851826c11209370923f17219a5f190e7acc806b4c73\": container with ID starting with 199e5efa4c1de85529465851826c11209370923f17219a5f190e7acc806b4c73 not found: ID does not exist" containerID="199e5efa4c1de85529465851826c11209370923f17219a5f190e7acc806b4c73" Dec 04 04:31:18 crc kubenswrapper[4806]: I1204 04:31:18.169072 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"199e5efa4c1de85529465851826c11209370923f17219a5f190e7acc806b4c73"} err="failed to get container status \"199e5efa4c1de85529465851826c11209370923f17219a5f190e7acc806b4c73\": rpc error: code = NotFound desc = could not find container \"199e5efa4c1de85529465851826c11209370923f17219a5f190e7acc806b4c73\": container with ID starting with 199e5efa4c1de85529465851826c11209370923f17219a5f190e7acc806b4c73 not found: ID does not exist" Dec 04 04:31:19 crc kubenswrapper[4806]: I1204 04:31:19.044158 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" event={"ID":"b4d109a9-916a-41b6-9008-15291d44a0a4","Type":"ContainerStarted","Data":"31afdaab101b444f4b038c9b66b71faf0e7695df9cac2356abbdeb9a36d5029c"} Dec 04 04:31:19 crc kubenswrapper[4806]: I1204 04:31:19.044535 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" event={"ID":"b4d109a9-916a-41b6-9008-15291d44a0a4","Type":"ContainerStarted","Data":"7ead933d0b0851ad84191dae2f4c500d57ce85e00fbecb2734bf8b58b5a20c3b"} Dec 04 04:31:19 crc kubenswrapper[4806]: I1204 04:31:19.067876 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" podStartSLOduration=1.8580397400000002 podStartE2EDuration="2.067826644s" podCreationTimestamp="2025-12-04 04:31:17 +0000 UTC" firstStartedPulling="2025-12-04 04:31:18.16436725 +0000 UTC m=+2193.022880178" lastFinishedPulling="2025-12-04 04:31:18.374154154 +0000 UTC m=+2193.232667082" observedRunningTime="2025-12-04 04:31:19.061801125 +0000 UTC m=+2193.920314073" watchObservedRunningTime="2025-12-04 04:31:19.067826644 +0000 UTC m=+2193.926339572" Dec 04 04:31:19 crc kubenswrapper[4806]: I1204 04:31:19.439215 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb2d3782-54c2-428b-b056-1e5ebd82323e" path="/var/lib/kubelet/pods/fb2d3782-54c2-428b-b056-1e5ebd82323e/volumes" Dec 04 04:31:23 crc kubenswrapper[4806]: I1204 04:31:23.424045 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:31:23 crc kubenswrapper[4806]: E1204 04:31:23.425003 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:31:37 crc kubenswrapper[4806]: I1204 04:31:37.423811 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:31:37 crc kubenswrapper[4806]: E1204 04:31:37.424539 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:31:50 crc kubenswrapper[4806]: I1204 04:31:50.423678 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:31:50 crc kubenswrapper[4806]: E1204 04:31:50.424458 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:32:03 crc kubenswrapper[4806]: I1204 04:32:03.424268 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:32:03 crc kubenswrapper[4806]: E1204 04:32:03.425169 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:32:16 crc kubenswrapper[4806]: I1204 04:32:16.425229 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:32:16 crc kubenswrapper[4806]: E1204 04:32:16.426571 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.170840 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2wd4v"] Dec 04 04:32:21 crc kubenswrapper[4806]: E1204 04:32:21.171949 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb2d3782-54c2-428b-b056-1e5ebd82323e" containerName="extract-content" Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.171967 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb2d3782-54c2-428b-b056-1e5ebd82323e" containerName="extract-content" Dec 04 04:32:21 crc kubenswrapper[4806]: E1204 04:32:21.171988 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb2d3782-54c2-428b-b056-1e5ebd82323e" containerName="extract-utilities" Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.171995 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb2d3782-54c2-428b-b056-1e5ebd82323e" containerName="extract-utilities" Dec 04 04:32:21 crc kubenswrapper[4806]: E1204 04:32:21.172019 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb2d3782-54c2-428b-b056-1e5ebd82323e" containerName="registry-server" Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.172026 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb2d3782-54c2-428b-b056-1e5ebd82323e" containerName="registry-server" Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.172234 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb2d3782-54c2-428b-b056-1e5ebd82323e" containerName="registry-server" Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.174128 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.180744 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2wd4v"] Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.314340 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-utilities\") pod \"community-operators-2wd4v\" (UID: \"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad\") " pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.314425 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxvvz\" (UniqueName: \"kubernetes.io/projected/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-kube-api-access-gxvvz\") pod \"community-operators-2wd4v\" (UID: \"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad\") " pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.314450 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-catalog-content\") pod \"community-operators-2wd4v\" (UID: \"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad\") " pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.416544 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-utilities\") pod \"community-operators-2wd4v\" (UID: \"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad\") " pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.416639 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxvvz\" (UniqueName: \"kubernetes.io/projected/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-kube-api-access-gxvvz\") pod \"community-operators-2wd4v\" (UID: \"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad\") " pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.416661 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-catalog-content\") pod \"community-operators-2wd4v\" (UID: \"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad\") " pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.417254 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-catalog-content\") pod \"community-operators-2wd4v\" (UID: \"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad\") " pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.417675 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-utilities\") pod \"community-operators-2wd4v\" (UID: \"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad\") " pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.450272 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxvvz\" (UniqueName: \"kubernetes.io/projected/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-kube-api-access-gxvvz\") pod \"community-operators-2wd4v\" (UID: \"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad\") " pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:21 crc kubenswrapper[4806]: I1204 04:32:21.504230 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:22 crc kubenswrapper[4806]: I1204 04:32:22.061223 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2wd4v"] Dec 04 04:32:22 crc kubenswrapper[4806]: I1204 04:32:22.600449 4806 generic.go:334] "Generic (PLEG): container finished" podID="ad65fbe4-31c8-4ac2-89d1-5d48ce572bad" containerID="9e55f8c884b2989edd3ca0ee469eb7c7f28430d61237e89487cc543602e868f8" exitCode=0 Dec 04 04:32:22 crc kubenswrapper[4806]: I1204 04:32:22.602848 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wd4v" event={"ID":"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad","Type":"ContainerDied","Data":"9e55f8c884b2989edd3ca0ee469eb7c7f28430d61237e89487cc543602e868f8"} Dec 04 04:32:22 crc kubenswrapper[4806]: I1204 04:32:22.602998 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wd4v" event={"ID":"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad","Type":"ContainerStarted","Data":"e8da96f49850a04adbe198b338b832c15054f0a0f6248b20ddddacd010066e3f"} Dec 04 04:32:24 crc kubenswrapper[4806]: I1204 04:32:24.622061 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wd4v" event={"ID":"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad","Type":"ContainerStarted","Data":"614dcf21bdc5516248b2b41293f819d03f53794eff0861d9b4bfb07ec9f3fe9c"} Dec 04 04:32:25 crc kubenswrapper[4806]: I1204 04:32:25.631849 4806 generic.go:334] "Generic (PLEG): container finished" podID="ad65fbe4-31c8-4ac2-89d1-5d48ce572bad" containerID="614dcf21bdc5516248b2b41293f819d03f53794eff0861d9b4bfb07ec9f3fe9c" exitCode=0 Dec 04 04:32:25 crc kubenswrapper[4806]: I1204 04:32:25.632067 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wd4v" event={"ID":"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad","Type":"ContainerDied","Data":"614dcf21bdc5516248b2b41293f819d03f53794eff0861d9b4bfb07ec9f3fe9c"} Dec 04 04:32:26 crc kubenswrapper[4806]: I1204 04:32:26.644499 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wd4v" event={"ID":"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad","Type":"ContainerStarted","Data":"bc3087a62a7964755933b555cc73abb8e534799a9b773b8a09f27dd61bf1d9f5"} Dec 04 04:32:26 crc kubenswrapper[4806]: I1204 04:32:26.673563 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2wd4v" podStartSLOduration=2.113624202 podStartE2EDuration="5.673535855s" podCreationTimestamp="2025-12-04 04:32:21 +0000 UTC" firstStartedPulling="2025-12-04 04:32:22.602616634 +0000 UTC m=+2257.461129562" lastFinishedPulling="2025-12-04 04:32:26.162528287 +0000 UTC m=+2261.021041215" observedRunningTime="2025-12-04 04:32:26.667044471 +0000 UTC m=+2261.525557389" watchObservedRunningTime="2025-12-04 04:32:26.673535855 +0000 UTC m=+2261.532048803" Dec 04 04:32:27 crc kubenswrapper[4806]: I1204 04:32:27.423786 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:32:27 crc kubenswrapper[4806]: E1204 04:32:27.424126 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:32:28 crc kubenswrapper[4806]: I1204 04:32:28.653704 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qjbnw"] Dec 04 04:32:28 crc kubenswrapper[4806]: I1204 04:32:28.695306 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:28 crc kubenswrapper[4806]: I1204 04:32:28.733421 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qjbnw"] Dec 04 04:32:28 crc kubenswrapper[4806]: I1204 04:32:28.781569 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncjdv\" (UniqueName: \"kubernetes.io/projected/a2e486e9-067d-4617-b2ed-c00497f5a9a2-kube-api-access-ncjdv\") pod \"redhat-marketplace-qjbnw\" (UID: \"a2e486e9-067d-4617-b2ed-c00497f5a9a2\") " pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:28 crc kubenswrapper[4806]: I1204 04:32:28.781697 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2e486e9-067d-4617-b2ed-c00497f5a9a2-catalog-content\") pod \"redhat-marketplace-qjbnw\" (UID: \"a2e486e9-067d-4617-b2ed-c00497f5a9a2\") " pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:28 crc kubenswrapper[4806]: I1204 04:32:28.781841 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2e486e9-067d-4617-b2ed-c00497f5a9a2-utilities\") pod \"redhat-marketplace-qjbnw\" (UID: \"a2e486e9-067d-4617-b2ed-c00497f5a9a2\") " pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:28 crc kubenswrapper[4806]: I1204 04:32:28.884075 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2e486e9-067d-4617-b2ed-c00497f5a9a2-catalog-content\") pod \"redhat-marketplace-qjbnw\" (UID: \"a2e486e9-067d-4617-b2ed-c00497f5a9a2\") " pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:28 crc kubenswrapper[4806]: I1204 04:32:28.884272 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2e486e9-067d-4617-b2ed-c00497f5a9a2-utilities\") pod \"redhat-marketplace-qjbnw\" (UID: \"a2e486e9-067d-4617-b2ed-c00497f5a9a2\") " pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:28 crc kubenswrapper[4806]: I1204 04:32:28.884370 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncjdv\" (UniqueName: \"kubernetes.io/projected/a2e486e9-067d-4617-b2ed-c00497f5a9a2-kube-api-access-ncjdv\") pod \"redhat-marketplace-qjbnw\" (UID: \"a2e486e9-067d-4617-b2ed-c00497f5a9a2\") " pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:28 crc kubenswrapper[4806]: I1204 04:32:28.884985 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2e486e9-067d-4617-b2ed-c00497f5a9a2-catalog-content\") pod \"redhat-marketplace-qjbnw\" (UID: \"a2e486e9-067d-4617-b2ed-c00497f5a9a2\") " pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:28 crc kubenswrapper[4806]: I1204 04:32:28.885060 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2e486e9-067d-4617-b2ed-c00497f5a9a2-utilities\") pod \"redhat-marketplace-qjbnw\" (UID: \"a2e486e9-067d-4617-b2ed-c00497f5a9a2\") " pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:28 crc kubenswrapper[4806]: I1204 04:32:28.907861 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncjdv\" (UniqueName: \"kubernetes.io/projected/a2e486e9-067d-4617-b2ed-c00497f5a9a2-kube-api-access-ncjdv\") pod \"redhat-marketplace-qjbnw\" (UID: \"a2e486e9-067d-4617-b2ed-c00497f5a9a2\") " pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:29 crc kubenswrapper[4806]: I1204 04:32:29.049151 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:29 crc kubenswrapper[4806]: I1204 04:32:29.611208 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qjbnw"] Dec 04 04:32:29 crc kubenswrapper[4806]: I1204 04:32:29.701937 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qjbnw" event={"ID":"a2e486e9-067d-4617-b2ed-c00497f5a9a2","Type":"ContainerStarted","Data":"2161fccdbdf9de7e95c2e8887ba41ff143a2ad5445830dd50f7480630b24e18a"} Dec 04 04:32:30 crc kubenswrapper[4806]: I1204 04:32:30.712592 4806 generic.go:334] "Generic (PLEG): container finished" podID="a2e486e9-067d-4617-b2ed-c00497f5a9a2" containerID="523b5068eb2d076b3ed14625ac90f0bd57d396ff524f1ed4b476a3a2620611f9" exitCode=0 Dec 04 04:32:30 crc kubenswrapper[4806]: I1204 04:32:30.712640 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qjbnw" event={"ID":"a2e486e9-067d-4617-b2ed-c00497f5a9a2","Type":"ContainerDied","Data":"523b5068eb2d076b3ed14625ac90f0bd57d396ff524f1ed4b476a3a2620611f9"} Dec 04 04:32:31 crc kubenswrapper[4806]: I1204 04:32:31.504609 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:31 crc kubenswrapper[4806]: I1204 04:32:31.504916 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:31 crc kubenswrapper[4806]: I1204 04:32:31.578847 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:31 crc kubenswrapper[4806]: I1204 04:32:31.722489 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qjbnw" event={"ID":"a2e486e9-067d-4617-b2ed-c00497f5a9a2","Type":"ContainerStarted","Data":"939199cfc107d127ee374cba2be709969f94e9ea3177316b2210487a60e1d74e"} Dec 04 04:32:31 crc kubenswrapper[4806]: I1204 04:32:31.788150 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:32 crc kubenswrapper[4806]: I1204 04:32:32.732626 4806 generic.go:334] "Generic (PLEG): container finished" podID="a2e486e9-067d-4617-b2ed-c00497f5a9a2" containerID="939199cfc107d127ee374cba2be709969f94e9ea3177316b2210487a60e1d74e" exitCode=0 Dec 04 04:32:32 crc kubenswrapper[4806]: I1204 04:32:32.732700 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qjbnw" event={"ID":"a2e486e9-067d-4617-b2ed-c00497f5a9a2","Type":"ContainerDied","Data":"939199cfc107d127ee374cba2be709969f94e9ea3177316b2210487a60e1d74e"} Dec 04 04:32:33 crc kubenswrapper[4806]: I1204 04:32:33.746157 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qjbnw" event={"ID":"a2e486e9-067d-4617-b2ed-c00497f5a9a2","Type":"ContainerStarted","Data":"db25c735d565ce2408c7c199283e15e1fdac3acc87034e7cf95dac42202da3e4"} Dec 04 04:32:33 crc kubenswrapper[4806]: I1204 04:32:33.769390 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qjbnw" podStartSLOduration=3.256334098 podStartE2EDuration="5.769365876s" podCreationTimestamp="2025-12-04 04:32:28 +0000 UTC" firstStartedPulling="2025-12-04 04:32:30.714955308 +0000 UTC m=+2265.573468236" lastFinishedPulling="2025-12-04 04:32:33.227987086 +0000 UTC m=+2268.086500014" observedRunningTime="2025-12-04 04:32:33.768258782 +0000 UTC m=+2268.626771720" watchObservedRunningTime="2025-12-04 04:32:33.769365876 +0000 UTC m=+2268.627878804" Dec 04 04:32:33 crc kubenswrapper[4806]: I1204 04:32:33.941642 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2wd4v"] Dec 04 04:32:34 crc kubenswrapper[4806]: I1204 04:32:34.755243 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2wd4v" podUID="ad65fbe4-31c8-4ac2-89d1-5d48ce572bad" containerName="registry-server" containerID="cri-o://bc3087a62a7964755933b555cc73abb8e534799a9b773b8a09f27dd61bf1d9f5" gracePeriod=2 Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.205230 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.321650 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxvvz\" (UniqueName: \"kubernetes.io/projected/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-kube-api-access-gxvvz\") pod \"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad\" (UID: \"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad\") " Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.321774 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-catalog-content\") pod \"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad\" (UID: \"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad\") " Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.321920 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-utilities\") pod \"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad\" (UID: \"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad\") " Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.322750 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-utilities" (OuterVolumeSpecName: "utilities") pod "ad65fbe4-31c8-4ac2-89d1-5d48ce572bad" (UID: "ad65fbe4-31c8-4ac2-89d1-5d48ce572bad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.336837 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-kube-api-access-gxvvz" (OuterVolumeSpecName: "kube-api-access-gxvvz") pod "ad65fbe4-31c8-4ac2-89d1-5d48ce572bad" (UID: "ad65fbe4-31c8-4ac2-89d1-5d48ce572bad"). InnerVolumeSpecName "kube-api-access-gxvvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.379143 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad65fbe4-31c8-4ac2-89d1-5d48ce572bad" (UID: "ad65fbe4-31c8-4ac2-89d1-5d48ce572bad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.441389 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.441439 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxvvz\" (UniqueName: \"kubernetes.io/projected/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-kube-api-access-gxvvz\") on node \"crc\" DevicePath \"\"" Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.441450 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.768300 4806 generic.go:334] "Generic (PLEG): container finished" podID="ad65fbe4-31c8-4ac2-89d1-5d48ce572bad" containerID="bc3087a62a7964755933b555cc73abb8e534799a9b773b8a09f27dd61bf1d9f5" exitCode=0 Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.768371 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2wd4v" Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.768411 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wd4v" event={"ID":"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad","Type":"ContainerDied","Data":"bc3087a62a7964755933b555cc73abb8e534799a9b773b8a09f27dd61bf1d9f5"} Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.768846 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wd4v" event={"ID":"ad65fbe4-31c8-4ac2-89d1-5d48ce572bad","Type":"ContainerDied","Data":"e8da96f49850a04adbe198b338b832c15054f0a0f6248b20ddddacd010066e3f"} Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.768878 4806 scope.go:117] "RemoveContainer" containerID="bc3087a62a7964755933b555cc73abb8e534799a9b773b8a09f27dd61bf1d9f5" Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.798058 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2wd4v"] Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.803392 4806 scope.go:117] "RemoveContainer" containerID="614dcf21bdc5516248b2b41293f819d03f53794eff0861d9b4bfb07ec9f3fe9c" Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.807856 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2wd4v"] Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.840985 4806 scope.go:117] "RemoveContainer" containerID="9e55f8c884b2989edd3ca0ee469eb7c7f28430d61237e89487cc543602e868f8" Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.888259 4806 scope.go:117] "RemoveContainer" containerID="bc3087a62a7964755933b555cc73abb8e534799a9b773b8a09f27dd61bf1d9f5" Dec 04 04:32:35 crc kubenswrapper[4806]: E1204 04:32:35.888831 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc3087a62a7964755933b555cc73abb8e534799a9b773b8a09f27dd61bf1d9f5\": container with ID starting with bc3087a62a7964755933b555cc73abb8e534799a9b773b8a09f27dd61bf1d9f5 not found: ID does not exist" containerID="bc3087a62a7964755933b555cc73abb8e534799a9b773b8a09f27dd61bf1d9f5" Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.888872 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc3087a62a7964755933b555cc73abb8e534799a9b773b8a09f27dd61bf1d9f5"} err="failed to get container status \"bc3087a62a7964755933b555cc73abb8e534799a9b773b8a09f27dd61bf1d9f5\": rpc error: code = NotFound desc = could not find container \"bc3087a62a7964755933b555cc73abb8e534799a9b773b8a09f27dd61bf1d9f5\": container with ID starting with bc3087a62a7964755933b555cc73abb8e534799a9b773b8a09f27dd61bf1d9f5 not found: ID does not exist" Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.888902 4806 scope.go:117] "RemoveContainer" containerID="614dcf21bdc5516248b2b41293f819d03f53794eff0861d9b4bfb07ec9f3fe9c" Dec 04 04:32:35 crc kubenswrapper[4806]: E1204 04:32:35.889393 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"614dcf21bdc5516248b2b41293f819d03f53794eff0861d9b4bfb07ec9f3fe9c\": container with ID starting with 614dcf21bdc5516248b2b41293f819d03f53794eff0861d9b4bfb07ec9f3fe9c not found: ID does not exist" containerID="614dcf21bdc5516248b2b41293f819d03f53794eff0861d9b4bfb07ec9f3fe9c" Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.889446 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"614dcf21bdc5516248b2b41293f819d03f53794eff0861d9b4bfb07ec9f3fe9c"} err="failed to get container status \"614dcf21bdc5516248b2b41293f819d03f53794eff0861d9b4bfb07ec9f3fe9c\": rpc error: code = NotFound desc = could not find container \"614dcf21bdc5516248b2b41293f819d03f53794eff0861d9b4bfb07ec9f3fe9c\": container with ID starting with 614dcf21bdc5516248b2b41293f819d03f53794eff0861d9b4bfb07ec9f3fe9c not found: ID does not exist" Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.889478 4806 scope.go:117] "RemoveContainer" containerID="9e55f8c884b2989edd3ca0ee469eb7c7f28430d61237e89487cc543602e868f8" Dec 04 04:32:35 crc kubenswrapper[4806]: E1204 04:32:35.889992 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e55f8c884b2989edd3ca0ee469eb7c7f28430d61237e89487cc543602e868f8\": container with ID starting with 9e55f8c884b2989edd3ca0ee469eb7c7f28430d61237e89487cc543602e868f8 not found: ID does not exist" containerID="9e55f8c884b2989edd3ca0ee469eb7c7f28430d61237e89487cc543602e868f8" Dec 04 04:32:35 crc kubenswrapper[4806]: I1204 04:32:35.890016 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e55f8c884b2989edd3ca0ee469eb7c7f28430d61237e89487cc543602e868f8"} err="failed to get container status \"9e55f8c884b2989edd3ca0ee469eb7c7f28430d61237e89487cc543602e868f8\": rpc error: code = NotFound desc = could not find container \"9e55f8c884b2989edd3ca0ee469eb7c7f28430d61237e89487cc543602e868f8\": container with ID starting with 9e55f8c884b2989edd3ca0ee469eb7c7f28430d61237e89487cc543602e868f8 not found: ID does not exist" Dec 04 04:32:37 crc kubenswrapper[4806]: I1204 04:32:37.435097 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad65fbe4-31c8-4ac2-89d1-5d48ce572bad" path="/var/lib/kubelet/pods/ad65fbe4-31c8-4ac2-89d1-5d48ce572bad/volumes" Dec 04 04:32:39 crc kubenswrapper[4806]: I1204 04:32:39.049575 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:39 crc kubenswrapper[4806]: I1204 04:32:39.050826 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:39 crc kubenswrapper[4806]: I1204 04:32:39.100599 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:39 crc kubenswrapper[4806]: I1204 04:32:39.519699 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:32:39 crc kubenswrapper[4806]: E1204 04:32:39.520018 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:32:39 crc kubenswrapper[4806]: I1204 04:32:39.800780 4806 generic.go:334] "Generic (PLEG): container finished" podID="b4d109a9-916a-41b6-9008-15291d44a0a4" containerID="31afdaab101b444f4b038c9b66b71faf0e7695df9cac2356abbdeb9a36d5029c" exitCode=0 Dec 04 04:32:39 crc kubenswrapper[4806]: I1204 04:32:39.800881 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" event={"ID":"b4d109a9-916a-41b6-9008-15291d44a0a4","Type":"ContainerDied","Data":"31afdaab101b444f4b038c9b66b71faf0e7695df9cac2356abbdeb9a36d5029c"} Dec 04 04:32:39 crc kubenswrapper[4806]: I1204 04:32:39.854024 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:39 crc kubenswrapper[4806]: I1204 04:32:39.937387 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qjbnw"] Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.261367 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.348129 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-ssh-key\") pod \"b4d109a9-916a-41b6-9008-15291d44a0a4\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.348498 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-ovn-combined-ca-bundle\") pod \"b4d109a9-916a-41b6-9008-15291d44a0a4\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.348564 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s94l\" (UniqueName: \"kubernetes.io/projected/b4d109a9-916a-41b6-9008-15291d44a0a4-kube-api-access-5s94l\") pod \"b4d109a9-916a-41b6-9008-15291d44a0a4\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.348714 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b4d109a9-916a-41b6-9008-15291d44a0a4-ovncontroller-config-0\") pod \"b4d109a9-916a-41b6-9008-15291d44a0a4\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.348771 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-inventory\") pod \"b4d109a9-916a-41b6-9008-15291d44a0a4\" (UID: \"b4d109a9-916a-41b6-9008-15291d44a0a4\") " Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.355354 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b4d109a9-916a-41b6-9008-15291d44a0a4" (UID: "b4d109a9-916a-41b6-9008-15291d44a0a4"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.368283 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4d109a9-916a-41b6-9008-15291d44a0a4-kube-api-access-5s94l" (OuterVolumeSpecName: "kube-api-access-5s94l") pod "b4d109a9-916a-41b6-9008-15291d44a0a4" (UID: "b4d109a9-916a-41b6-9008-15291d44a0a4"). InnerVolumeSpecName "kube-api-access-5s94l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.386356 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b4d109a9-916a-41b6-9008-15291d44a0a4" (UID: "b4d109a9-916a-41b6-9008-15291d44a0a4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.388788 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-inventory" (OuterVolumeSpecName: "inventory") pod "b4d109a9-916a-41b6-9008-15291d44a0a4" (UID: "b4d109a9-916a-41b6-9008-15291d44a0a4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.408913 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4d109a9-916a-41b6-9008-15291d44a0a4-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "b4d109a9-916a-41b6-9008-15291d44a0a4" (UID: "b4d109a9-916a-41b6-9008-15291d44a0a4"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.451499 4806 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b4d109a9-916a-41b6-9008-15291d44a0a4-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.451554 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.451590 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.451603 4806 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4d109a9-916a-41b6-9008-15291d44a0a4-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.451614 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s94l\" (UniqueName: \"kubernetes.io/projected/b4d109a9-916a-41b6-9008-15291d44a0a4-kube-api-access-5s94l\") on node \"crc\" DevicePath \"\"" Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.818069 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qjbnw" podUID="a2e486e9-067d-4617-b2ed-c00497f5a9a2" containerName="registry-server" containerID="cri-o://db25c735d565ce2408c7c199283e15e1fdac3acc87034e7cf95dac42202da3e4" gracePeriod=2 Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.818377 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" event={"ID":"b4d109a9-916a-41b6-9008-15291d44a0a4","Type":"ContainerDied","Data":"7ead933d0b0851ad84191dae2f4c500d57ce85e00fbecb2734bf8b58b5a20c3b"} Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.818420 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ead933d0b0851ad84191dae2f4c500d57ce85e00fbecb2734bf8b58b5a20c3b" Dec 04 04:32:41 crc kubenswrapper[4806]: I1204 04:32:41.818422 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-mbhj2" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.021743 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq"] Dec 04 04:32:42 crc kubenswrapper[4806]: E1204 04:32:42.022408 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad65fbe4-31c8-4ac2-89d1-5d48ce572bad" containerName="extract-utilities" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.022426 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad65fbe4-31c8-4ac2-89d1-5d48ce572bad" containerName="extract-utilities" Dec 04 04:32:42 crc kubenswrapper[4806]: E1204 04:32:42.022443 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4d109a9-916a-41b6-9008-15291d44a0a4" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.022451 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4d109a9-916a-41b6-9008-15291d44a0a4" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 04 04:32:42 crc kubenswrapper[4806]: E1204 04:32:42.022486 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad65fbe4-31c8-4ac2-89d1-5d48ce572bad" containerName="extract-content" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.022494 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad65fbe4-31c8-4ac2-89d1-5d48ce572bad" containerName="extract-content" Dec 04 04:32:42 crc kubenswrapper[4806]: E1204 04:32:42.022503 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad65fbe4-31c8-4ac2-89d1-5d48ce572bad" containerName="registry-server" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.022510 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad65fbe4-31c8-4ac2-89d1-5d48ce572bad" containerName="registry-server" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.022685 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4d109a9-916a-41b6-9008-15291d44a0a4" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.022702 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad65fbe4-31c8-4ac2-89d1-5d48ce572bad" containerName="registry-server" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.024970 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.027466 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.027503 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.027576 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-q8mqk" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.027585 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.027615 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.027755 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.050530 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq"] Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.169522 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.169589 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.169653 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.169804 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bln9c\" (UniqueName: \"kubernetes.io/projected/4b974095-c38a-43c6-a6ac-67497d4f509f-kube-api-access-bln9c\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.169832 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.169852 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.272152 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.272220 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.272281 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.272333 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bln9c\" (UniqueName: \"kubernetes.io/projected/4b974095-c38a-43c6-a6ac-67497d4f509f-kube-api-access-bln9c\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.272365 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.272390 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.278311 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.279081 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.279199 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.280473 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.281151 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.290756 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bln9c\" (UniqueName: \"kubernetes.io/projected/4b974095-c38a-43c6-a6ac-67497d4f509f-kube-api-access-bln9c\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.348654 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.352169 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.499685 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2e486e9-067d-4617-b2ed-c00497f5a9a2-catalog-content\") pod \"a2e486e9-067d-4617-b2ed-c00497f5a9a2\" (UID: \"a2e486e9-067d-4617-b2ed-c00497f5a9a2\") " Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.500041 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2e486e9-067d-4617-b2ed-c00497f5a9a2-utilities\") pod \"a2e486e9-067d-4617-b2ed-c00497f5a9a2\" (UID: \"a2e486e9-067d-4617-b2ed-c00497f5a9a2\") " Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.500290 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncjdv\" (UniqueName: \"kubernetes.io/projected/a2e486e9-067d-4617-b2ed-c00497f5a9a2-kube-api-access-ncjdv\") pod \"a2e486e9-067d-4617-b2ed-c00497f5a9a2\" (UID: \"a2e486e9-067d-4617-b2ed-c00497f5a9a2\") " Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.500780 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2e486e9-067d-4617-b2ed-c00497f5a9a2-utilities" (OuterVolumeSpecName: "utilities") pod "a2e486e9-067d-4617-b2ed-c00497f5a9a2" (UID: "a2e486e9-067d-4617-b2ed-c00497f5a9a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.504583 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2e486e9-067d-4617-b2ed-c00497f5a9a2-kube-api-access-ncjdv" (OuterVolumeSpecName: "kube-api-access-ncjdv") pod "a2e486e9-067d-4617-b2ed-c00497f5a9a2" (UID: "a2e486e9-067d-4617-b2ed-c00497f5a9a2"). InnerVolumeSpecName "kube-api-access-ncjdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.542156 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2e486e9-067d-4617-b2ed-c00497f5a9a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2e486e9-067d-4617-b2ed-c00497f5a9a2" (UID: "a2e486e9-067d-4617-b2ed-c00497f5a9a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.603349 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2e486e9-067d-4617-b2ed-c00497f5a9a2-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.603381 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncjdv\" (UniqueName: \"kubernetes.io/projected/a2e486e9-067d-4617-b2ed-c00497f5a9a2-kube-api-access-ncjdv\") on node \"crc\" DevicePath \"\"" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.603393 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2e486e9-067d-4617-b2ed-c00497f5a9a2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.828854 4806 generic.go:334] "Generic (PLEG): container finished" podID="a2e486e9-067d-4617-b2ed-c00497f5a9a2" containerID="db25c735d565ce2408c7c199283e15e1fdac3acc87034e7cf95dac42202da3e4" exitCode=0 Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.828958 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qjbnw" event={"ID":"a2e486e9-067d-4617-b2ed-c00497f5a9a2","Type":"ContainerDied","Data":"db25c735d565ce2408c7c199283e15e1fdac3acc87034e7cf95dac42202da3e4"} Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.828970 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qjbnw" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.829263 4806 scope.go:117] "RemoveContainer" containerID="db25c735d565ce2408c7c199283e15e1fdac3acc87034e7cf95dac42202da3e4" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.829246 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qjbnw" event={"ID":"a2e486e9-067d-4617-b2ed-c00497f5a9a2","Type":"ContainerDied","Data":"2161fccdbdf9de7e95c2e8887ba41ff143a2ad5445830dd50f7480630b24e18a"} Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.853664 4806 scope.go:117] "RemoveContainer" containerID="939199cfc107d127ee374cba2be709969f94e9ea3177316b2210487a60e1d74e" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.886799 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qjbnw"] Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.891201 4806 scope.go:117] "RemoveContainer" containerID="523b5068eb2d076b3ed14625ac90f0bd57d396ff524f1ed4b476a3a2620611f9" Dec 04 04:32:42 crc kubenswrapper[4806]: I1204 04:32:42.909457 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qjbnw"] Dec 04 04:32:43 crc kubenswrapper[4806]: I1204 04:32:43.025033 4806 scope.go:117] "RemoveContainer" containerID="db25c735d565ce2408c7c199283e15e1fdac3acc87034e7cf95dac42202da3e4" Dec 04 04:32:43 crc kubenswrapper[4806]: E1204 04:32:43.029405 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db25c735d565ce2408c7c199283e15e1fdac3acc87034e7cf95dac42202da3e4\": container with ID starting with db25c735d565ce2408c7c199283e15e1fdac3acc87034e7cf95dac42202da3e4 not found: ID does not exist" containerID="db25c735d565ce2408c7c199283e15e1fdac3acc87034e7cf95dac42202da3e4" Dec 04 04:32:43 crc kubenswrapper[4806]: I1204 04:32:43.029483 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db25c735d565ce2408c7c199283e15e1fdac3acc87034e7cf95dac42202da3e4"} err="failed to get container status \"db25c735d565ce2408c7c199283e15e1fdac3acc87034e7cf95dac42202da3e4\": rpc error: code = NotFound desc = could not find container \"db25c735d565ce2408c7c199283e15e1fdac3acc87034e7cf95dac42202da3e4\": container with ID starting with db25c735d565ce2408c7c199283e15e1fdac3acc87034e7cf95dac42202da3e4 not found: ID does not exist" Dec 04 04:32:43 crc kubenswrapper[4806]: I1204 04:32:43.029537 4806 scope.go:117] "RemoveContainer" containerID="939199cfc107d127ee374cba2be709969f94e9ea3177316b2210487a60e1d74e" Dec 04 04:32:43 crc kubenswrapper[4806]: E1204 04:32:43.033296 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"939199cfc107d127ee374cba2be709969f94e9ea3177316b2210487a60e1d74e\": container with ID starting with 939199cfc107d127ee374cba2be709969f94e9ea3177316b2210487a60e1d74e not found: ID does not exist" containerID="939199cfc107d127ee374cba2be709969f94e9ea3177316b2210487a60e1d74e" Dec 04 04:32:43 crc kubenswrapper[4806]: I1204 04:32:43.033343 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"939199cfc107d127ee374cba2be709969f94e9ea3177316b2210487a60e1d74e"} err="failed to get container status \"939199cfc107d127ee374cba2be709969f94e9ea3177316b2210487a60e1d74e\": rpc error: code = NotFound desc = could not find container \"939199cfc107d127ee374cba2be709969f94e9ea3177316b2210487a60e1d74e\": container with ID starting with 939199cfc107d127ee374cba2be709969f94e9ea3177316b2210487a60e1d74e not found: ID does not exist" Dec 04 04:32:43 crc kubenswrapper[4806]: I1204 04:32:43.033369 4806 scope.go:117] "RemoveContainer" containerID="523b5068eb2d076b3ed14625ac90f0bd57d396ff524f1ed4b476a3a2620611f9" Dec 04 04:32:43 crc kubenswrapper[4806]: E1204 04:32:43.039135 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"523b5068eb2d076b3ed14625ac90f0bd57d396ff524f1ed4b476a3a2620611f9\": container with ID starting with 523b5068eb2d076b3ed14625ac90f0bd57d396ff524f1ed4b476a3a2620611f9 not found: ID does not exist" containerID="523b5068eb2d076b3ed14625ac90f0bd57d396ff524f1ed4b476a3a2620611f9" Dec 04 04:32:43 crc kubenswrapper[4806]: I1204 04:32:43.039202 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"523b5068eb2d076b3ed14625ac90f0bd57d396ff524f1ed4b476a3a2620611f9"} err="failed to get container status \"523b5068eb2d076b3ed14625ac90f0bd57d396ff524f1ed4b476a3a2620611f9\": rpc error: code = NotFound desc = could not find container \"523b5068eb2d076b3ed14625ac90f0bd57d396ff524f1ed4b476a3a2620611f9\": container with ID starting with 523b5068eb2d076b3ed14625ac90f0bd57d396ff524f1ed4b476a3a2620611f9 not found: ID does not exist" Dec 04 04:32:43 crc kubenswrapper[4806]: I1204 04:32:43.056656 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq"] Dec 04 04:32:43 crc kubenswrapper[4806]: I1204 04:32:43.437871 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2e486e9-067d-4617-b2ed-c00497f5a9a2" path="/var/lib/kubelet/pods/a2e486e9-067d-4617-b2ed-c00497f5a9a2/volumes" Dec 04 04:32:43 crc kubenswrapper[4806]: I1204 04:32:43.840232 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" event={"ID":"4b974095-c38a-43c6-a6ac-67497d4f509f","Type":"ContainerStarted","Data":"8347f7761205cd63151da9cde48787b38edfbda35526b34abcc5660a449158a2"} Dec 04 04:32:43 crc kubenswrapper[4806]: I1204 04:32:43.840578 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" event={"ID":"4b974095-c38a-43c6-a6ac-67497d4f509f","Type":"ContainerStarted","Data":"670026b4015130ff612a03a2c9b2ac246014c8ac55d79b1a262ed385a5075245"} Dec 04 04:32:43 crc kubenswrapper[4806]: I1204 04:32:43.867205 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" podStartSLOduration=2.673382238 podStartE2EDuration="2.867184469s" podCreationTimestamp="2025-12-04 04:32:41 +0000 UTC" firstStartedPulling="2025-12-04 04:32:43.177917769 +0000 UTC m=+2278.036430697" lastFinishedPulling="2025-12-04 04:32:43.37172 +0000 UTC m=+2278.230232928" observedRunningTime="2025-12-04 04:32:43.857461925 +0000 UTC m=+2278.715974853" watchObservedRunningTime="2025-12-04 04:32:43.867184469 +0000 UTC m=+2278.725697397" Dec 04 04:32:52 crc kubenswrapper[4806]: I1204 04:32:52.424481 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:32:52 crc kubenswrapper[4806]: E1204 04:32:52.425262 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:33:03 crc kubenswrapper[4806]: I1204 04:33:03.424353 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:33:03 crc kubenswrapper[4806]: E1204 04:33:03.425201 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:33:16 crc kubenswrapper[4806]: I1204 04:33:16.423465 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:33:16 crc kubenswrapper[4806]: E1204 04:33:16.424386 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:33:27 crc kubenswrapper[4806]: I1204 04:33:27.424694 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:33:27 crc kubenswrapper[4806]: E1204 04:33:27.425333 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:33:40 crc kubenswrapper[4806]: I1204 04:33:40.423866 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:33:40 crc kubenswrapper[4806]: E1204 04:33:40.424750 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:33:43 crc kubenswrapper[4806]: I1204 04:33:43.360652 4806 generic.go:334] "Generic (PLEG): container finished" podID="4b974095-c38a-43c6-a6ac-67497d4f509f" containerID="8347f7761205cd63151da9cde48787b38edfbda35526b34abcc5660a449158a2" exitCode=0 Dec 04 04:33:43 crc kubenswrapper[4806]: I1204 04:33:43.361001 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" event={"ID":"4b974095-c38a-43c6-a6ac-67497d4f509f","Type":"ContainerDied","Data":"8347f7761205cd63151da9cde48787b38edfbda35526b34abcc5660a449158a2"} Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.802705 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.892247 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-nova-metadata-neutron-config-0\") pod \"4b974095-c38a-43c6-a6ac-67497d4f509f\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.892388 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bln9c\" (UniqueName: \"kubernetes.io/projected/4b974095-c38a-43c6-a6ac-67497d4f509f-kube-api-access-bln9c\") pod \"4b974095-c38a-43c6-a6ac-67497d4f509f\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.892429 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-ssh-key\") pod \"4b974095-c38a-43c6-a6ac-67497d4f509f\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.892475 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-neutron-ovn-metadata-agent-neutron-config-0\") pod \"4b974095-c38a-43c6-a6ac-67497d4f509f\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.892511 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-inventory\") pod \"4b974095-c38a-43c6-a6ac-67497d4f509f\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.892628 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-neutron-metadata-combined-ca-bundle\") pod \"4b974095-c38a-43c6-a6ac-67497d4f509f\" (UID: \"4b974095-c38a-43c6-a6ac-67497d4f509f\") " Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.929147 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4b974095-c38a-43c6-a6ac-67497d4f509f" (UID: "4b974095-c38a-43c6-a6ac-67497d4f509f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.929227 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "4b974095-c38a-43c6-a6ac-67497d4f509f" (UID: "4b974095-c38a-43c6-a6ac-67497d4f509f"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.929254 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b974095-c38a-43c6-a6ac-67497d4f509f-kube-api-access-bln9c" (OuterVolumeSpecName: "kube-api-access-bln9c") pod "4b974095-c38a-43c6-a6ac-67497d4f509f" (UID: "4b974095-c38a-43c6-a6ac-67497d4f509f"). InnerVolumeSpecName "kube-api-access-bln9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.935133 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "4b974095-c38a-43c6-a6ac-67497d4f509f" (UID: "4b974095-c38a-43c6-a6ac-67497d4f509f"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.957483 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "4b974095-c38a-43c6-a6ac-67497d4f509f" (UID: "4b974095-c38a-43c6-a6ac-67497d4f509f"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.958840 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-inventory" (OuterVolumeSpecName: "inventory") pod "4b974095-c38a-43c6-a6ac-67497d4f509f" (UID: "4b974095-c38a-43c6-a6ac-67497d4f509f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.995840 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bln9c\" (UniqueName: \"kubernetes.io/projected/4b974095-c38a-43c6-a6ac-67497d4f509f-kube-api-access-bln9c\") on node \"crc\" DevicePath \"\"" Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.995907 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.995947 4806 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.995964 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.995978 4806 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:33:44 crc kubenswrapper[4806]: I1204 04:33:44.995995 4806 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b974095-c38a-43c6-a6ac-67497d4f509f-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.380456 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" event={"ID":"4b974095-c38a-43c6-a6ac-67497d4f509f","Type":"ContainerDied","Data":"670026b4015130ff612a03a2c9b2ac246014c8ac55d79b1a262ed385a5075245"} Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.380505 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="670026b4015130ff612a03a2c9b2ac246014c8ac55d79b1a262ed385a5075245" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.380529 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.499127 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2"] Dec 04 04:33:45 crc kubenswrapper[4806]: E1204 04:33:45.499675 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b974095-c38a-43c6-a6ac-67497d4f509f" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.499701 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b974095-c38a-43c6-a6ac-67497d4f509f" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 04 04:33:45 crc kubenswrapper[4806]: E1204 04:33:45.499728 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2e486e9-067d-4617-b2ed-c00497f5a9a2" containerName="extract-content" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.499738 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2e486e9-067d-4617-b2ed-c00497f5a9a2" containerName="extract-content" Dec 04 04:33:45 crc kubenswrapper[4806]: E1204 04:33:45.499754 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2e486e9-067d-4617-b2ed-c00497f5a9a2" containerName="extract-utilities" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.499763 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2e486e9-067d-4617-b2ed-c00497f5a9a2" containerName="extract-utilities" Dec 04 04:33:45 crc kubenswrapper[4806]: E1204 04:33:45.499774 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2e486e9-067d-4617-b2ed-c00497f5a9a2" containerName="registry-server" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.499783 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2e486e9-067d-4617-b2ed-c00497f5a9a2" containerName="registry-server" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.500072 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b974095-c38a-43c6-a6ac-67497d4f509f" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.500117 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2e486e9-067d-4617-b2ed-c00497f5a9a2" containerName="registry-server" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.501233 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.506462 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.506641 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-q8mqk" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.506786 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.506988 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.507584 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.508711 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2"] Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.607678 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.608464 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.608583 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.608719 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.608912 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95gpv\" (UniqueName: \"kubernetes.io/projected/daec7717-fd5a-47be-a766-1ef8bbabab14-kube-api-access-95gpv\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.710845 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.710913 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.711111 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.711210 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95gpv\" (UniqueName: \"kubernetes.io/projected/daec7717-fd5a-47be-a766-1ef8bbabab14-kube-api-access-95gpv\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.711240 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.717260 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.718318 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.718399 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.721422 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.730331 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95gpv\" (UniqueName: \"kubernetes.io/projected/daec7717-fd5a-47be-a766-1ef8bbabab14-kube-api-access-95gpv\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:45 crc kubenswrapper[4806]: I1204 04:33:45.819275 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:33:46 crc kubenswrapper[4806]: I1204 04:33:46.393797 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 04:33:46 crc kubenswrapper[4806]: I1204 04:33:46.397034 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2"] Dec 04 04:33:47 crc kubenswrapper[4806]: I1204 04:33:47.402515 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" event={"ID":"daec7717-fd5a-47be-a766-1ef8bbabab14","Type":"ContainerStarted","Data":"aeca461a45dc7237b7ac78ba275aa9de3e20b62731d136a0e81892a32fc8e2ba"} Dec 04 04:33:47 crc kubenswrapper[4806]: I1204 04:33:47.405488 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" event={"ID":"daec7717-fd5a-47be-a766-1ef8bbabab14","Type":"ContainerStarted","Data":"47b95613dcc0dc08564ebb4da595b667f06c5fd725ea8574f70f9314db76a59a"} Dec 04 04:33:51 crc kubenswrapper[4806]: I1204 04:33:51.424656 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:33:51 crc kubenswrapper[4806]: E1204 04:33:51.425518 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:34:06 crc kubenswrapper[4806]: I1204 04:34:06.423857 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:34:06 crc kubenswrapper[4806]: E1204 04:34:06.424589 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:34:17 crc kubenswrapper[4806]: I1204 04:34:17.423066 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:34:17 crc kubenswrapper[4806]: E1204 04:34:17.423848 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:34:31 crc kubenswrapper[4806]: I1204 04:34:31.424466 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:34:31 crc kubenswrapper[4806]: E1204 04:34:31.425075 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:34:43 crc kubenswrapper[4806]: I1204 04:34:43.423765 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:34:43 crc kubenswrapper[4806]: E1204 04:34:43.426076 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:34:57 crc kubenswrapper[4806]: I1204 04:34:57.423661 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:34:57 crc kubenswrapper[4806]: E1204 04:34:57.424351 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:35:11 crc kubenswrapper[4806]: I1204 04:35:11.424503 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:35:11 crc kubenswrapper[4806]: E1204 04:35:11.426238 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:35:22 crc kubenswrapper[4806]: I1204 04:35:22.424340 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:35:22 crc kubenswrapper[4806]: E1204 04:35:22.425201 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:35:36 crc kubenswrapper[4806]: I1204 04:35:36.424443 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:35:36 crc kubenswrapper[4806]: E1204 04:35:36.425300 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:35:47 crc kubenswrapper[4806]: I1204 04:35:47.425036 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:35:47 crc kubenswrapper[4806]: E1204 04:35:47.426965 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:36:00 crc kubenswrapper[4806]: I1204 04:36:00.423528 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:36:00 crc kubenswrapper[4806]: I1204 04:36:00.918362 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"cdb2e69904c8b363a34e509cfc2460a62051fc0403a906037069cecc07724904"} Dec 04 04:36:00 crc kubenswrapper[4806]: I1204 04:36:00.953889 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" podStartSLOduration=135.746251607 podStartE2EDuration="2m15.953865473s" podCreationTimestamp="2025-12-04 04:33:45 +0000 UTC" firstStartedPulling="2025-12-04 04:33:46.393590864 +0000 UTC m=+2341.252103792" lastFinishedPulling="2025-12-04 04:33:46.60120473 +0000 UTC m=+2341.459717658" observedRunningTime="2025-12-04 04:33:47.430382472 +0000 UTC m=+2342.288895400" watchObservedRunningTime="2025-12-04 04:36:00.953865473 +0000 UTC m=+2475.812378401" Dec 04 04:38:27 crc kubenswrapper[4806]: I1204 04:38:27.047795 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:38:27 crc kubenswrapper[4806]: I1204 04:38:27.049056 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:38:31 crc kubenswrapper[4806]: I1204 04:38:31.253062 4806 generic.go:334] "Generic (PLEG): container finished" podID="daec7717-fd5a-47be-a766-1ef8bbabab14" containerID="aeca461a45dc7237b7ac78ba275aa9de3e20b62731d136a0e81892a32fc8e2ba" exitCode=0 Dec 04 04:38:31 crc kubenswrapper[4806]: I1204 04:38:31.253116 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" event={"ID":"daec7717-fd5a-47be-a766-1ef8bbabab14","Type":"ContainerDied","Data":"aeca461a45dc7237b7ac78ba275aa9de3e20b62731d136a0e81892a32fc8e2ba"} Dec 04 04:38:32 crc kubenswrapper[4806]: I1204 04:38:32.717006 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:38:32 crc kubenswrapper[4806]: I1204 04:38:32.811131 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-inventory\") pod \"daec7717-fd5a-47be-a766-1ef8bbabab14\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " Dec 04 04:38:32 crc kubenswrapper[4806]: I1204 04:38:32.811182 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-libvirt-combined-ca-bundle\") pod \"daec7717-fd5a-47be-a766-1ef8bbabab14\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " Dec 04 04:38:32 crc kubenswrapper[4806]: I1204 04:38:32.811266 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-ssh-key\") pod \"daec7717-fd5a-47be-a766-1ef8bbabab14\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " Dec 04 04:38:32 crc kubenswrapper[4806]: I1204 04:38:32.811372 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-libvirt-secret-0\") pod \"daec7717-fd5a-47be-a766-1ef8bbabab14\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " Dec 04 04:38:32 crc kubenswrapper[4806]: I1204 04:38:32.811429 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95gpv\" (UniqueName: \"kubernetes.io/projected/daec7717-fd5a-47be-a766-1ef8bbabab14-kube-api-access-95gpv\") pod \"daec7717-fd5a-47be-a766-1ef8bbabab14\" (UID: \"daec7717-fd5a-47be-a766-1ef8bbabab14\") " Dec 04 04:38:32 crc kubenswrapper[4806]: I1204 04:38:32.821763 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daec7717-fd5a-47be-a766-1ef8bbabab14-kube-api-access-95gpv" (OuterVolumeSpecName: "kube-api-access-95gpv") pod "daec7717-fd5a-47be-a766-1ef8bbabab14" (UID: "daec7717-fd5a-47be-a766-1ef8bbabab14"). InnerVolumeSpecName "kube-api-access-95gpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:38:32 crc kubenswrapper[4806]: I1204 04:38:32.832557 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "daec7717-fd5a-47be-a766-1ef8bbabab14" (UID: "daec7717-fd5a-47be-a766-1ef8bbabab14"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:38:32 crc kubenswrapper[4806]: I1204 04:38:32.843208 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "daec7717-fd5a-47be-a766-1ef8bbabab14" (UID: "daec7717-fd5a-47be-a766-1ef8bbabab14"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:38:32 crc kubenswrapper[4806]: I1204 04:38:32.844144 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-inventory" (OuterVolumeSpecName: "inventory") pod "daec7717-fd5a-47be-a766-1ef8bbabab14" (UID: "daec7717-fd5a-47be-a766-1ef8bbabab14"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:38:32 crc kubenswrapper[4806]: I1204 04:38:32.846398 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "daec7717-fd5a-47be-a766-1ef8bbabab14" (UID: "daec7717-fd5a-47be-a766-1ef8bbabab14"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:38:32 crc kubenswrapper[4806]: I1204 04:38:32.913131 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95gpv\" (UniqueName: \"kubernetes.io/projected/daec7717-fd5a-47be-a766-1ef8bbabab14-kube-api-access-95gpv\") on node \"crc\" DevicePath \"\"" Dec 04 04:38:32 crc kubenswrapper[4806]: I1204 04:38:32.913161 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 04:38:32 crc kubenswrapper[4806]: I1204 04:38:32.913170 4806 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:38:32 crc kubenswrapper[4806]: I1204 04:38:32.913180 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:38:32 crc kubenswrapper[4806]: I1204 04:38:32.913191 4806 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/daec7717-fd5a-47be-a766-1ef8bbabab14-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.308479 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" event={"ID":"daec7717-fd5a-47be-a766-1ef8bbabab14","Type":"ContainerDied","Data":"47b95613dcc0dc08564ebb4da595b667f06c5fd725ea8574f70f9314db76a59a"} Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.308527 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47b95613dcc0dc08564ebb4da595b667f06c5fd725ea8574f70f9314db76a59a" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.308536 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.358983 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8"] Dec 04 04:38:33 crc kubenswrapper[4806]: E1204 04:38:33.359581 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daec7717-fd5a-47be-a766-1ef8bbabab14" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.359595 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="daec7717-fd5a-47be-a766-1ef8bbabab14" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.359783 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="daec7717-fd5a-47be-a766-1ef8bbabab14" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.360396 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.364270 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.364496 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.364561 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.364606 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.364786 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-q8mqk" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.365032 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.365218 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.379540 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8"] Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.423121 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.423190 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.423219 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.423313 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.423350 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.423369 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.423428 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.423447 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.423466 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrbb8\" (UniqueName: \"kubernetes.io/projected/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-kube-api-access-wrbb8\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.525542 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.525626 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.525646 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.525757 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.525789 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.525816 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrbb8\" (UniqueName: \"kubernetes.io/projected/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-kube-api-access-wrbb8\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.525843 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.525873 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.525953 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.527621 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.531739 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.531907 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.533383 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.533797 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.534043 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.534285 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.537379 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.545260 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrbb8\" (UniqueName: \"kubernetes.io/projected/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-kube-api-access-wrbb8\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l4hb8\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:33 crc kubenswrapper[4806]: I1204 04:38:33.677743 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:38:34 crc kubenswrapper[4806]: I1204 04:38:34.219777 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8"] Dec 04 04:38:34 crc kubenswrapper[4806]: W1204 04:38:34.234747 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21f3ccbb_669c_4055_b85c_f0c69fe61a8e.slice/crio-50e58dce49a9263258e2ae88322de0295cce329cddbd46116472df428ad005fd WatchSource:0}: Error finding container 50e58dce49a9263258e2ae88322de0295cce329cddbd46116472df428ad005fd: Status 404 returned error can't find the container with id 50e58dce49a9263258e2ae88322de0295cce329cddbd46116472df428ad005fd Dec 04 04:38:34 crc kubenswrapper[4806]: I1204 04:38:34.319493 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" event={"ID":"21f3ccbb-669c-4055-b85c-f0c69fe61a8e","Type":"ContainerStarted","Data":"50e58dce49a9263258e2ae88322de0295cce329cddbd46116472df428ad005fd"} Dec 04 04:38:35 crc kubenswrapper[4806]: I1204 04:38:35.336346 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" event={"ID":"21f3ccbb-669c-4055-b85c-f0c69fe61a8e","Type":"ContainerStarted","Data":"15c5843b9480b1fe84846753d32010e84d4b0b6f181d0964e77cc7db82d79ea4"} Dec 04 04:38:35 crc kubenswrapper[4806]: I1204 04:38:35.366437 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" podStartSLOduration=2.18300623 podStartE2EDuration="2.366410982s" podCreationTimestamp="2025-12-04 04:38:33 +0000 UTC" firstStartedPulling="2025-12-04 04:38:34.237682305 +0000 UTC m=+2629.096195233" lastFinishedPulling="2025-12-04 04:38:34.421087057 +0000 UTC m=+2629.279599985" observedRunningTime="2025-12-04 04:38:35.356007414 +0000 UTC m=+2630.214520342" watchObservedRunningTime="2025-12-04 04:38:35.366410982 +0000 UTC m=+2630.224923910" Dec 04 04:38:43 crc kubenswrapper[4806]: I1204 04:38:43.189316 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p52r9"] Dec 04 04:38:43 crc kubenswrapper[4806]: I1204 04:38:43.191953 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:38:43 crc kubenswrapper[4806]: I1204 04:38:43.202205 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p52r9"] Dec 04 04:38:43 crc kubenswrapper[4806]: I1204 04:38:43.230143 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b94f0115-4cdc-4682-a897-cd8e2f84ab27-catalog-content\") pod \"redhat-operators-p52r9\" (UID: \"b94f0115-4cdc-4682-a897-cd8e2f84ab27\") " pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:38:43 crc kubenswrapper[4806]: I1204 04:38:43.230346 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlssj\" (UniqueName: \"kubernetes.io/projected/b94f0115-4cdc-4682-a897-cd8e2f84ab27-kube-api-access-rlssj\") pod \"redhat-operators-p52r9\" (UID: \"b94f0115-4cdc-4682-a897-cd8e2f84ab27\") " pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:38:43 crc kubenswrapper[4806]: I1204 04:38:43.230405 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b94f0115-4cdc-4682-a897-cd8e2f84ab27-utilities\") pod \"redhat-operators-p52r9\" (UID: \"b94f0115-4cdc-4682-a897-cd8e2f84ab27\") " pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:38:43 crc kubenswrapper[4806]: I1204 04:38:43.332186 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlssj\" (UniqueName: \"kubernetes.io/projected/b94f0115-4cdc-4682-a897-cd8e2f84ab27-kube-api-access-rlssj\") pod \"redhat-operators-p52r9\" (UID: \"b94f0115-4cdc-4682-a897-cd8e2f84ab27\") " pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:38:43 crc kubenswrapper[4806]: I1204 04:38:43.332273 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b94f0115-4cdc-4682-a897-cd8e2f84ab27-utilities\") pod \"redhat-operators-p52r9\" (UID: \"b94f0115-4cdc-4682-a897-cd8e2f84ab27\") " pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:38:43 crc kubenswrapper[4806]: I1204 04:38:43.332359 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b94f0115-4cdc-4682-a897-cd8e2f84ab27-catalog-content\") pod \"redhat-operators-p52r9\" (UID: \"b94f0115-4cdc-4682-a897-cd8e2f84ab27\") " pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:38:43 crc kubenswrapper[4806]: I1204 04:38:43.332839 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b94f0115-4cdc-4682-a897-cd8e2f84ab27-utilities\") pod \"redhat-operators-p52r9\" (UID: \"b94f0115-4cdc-4682-a897-cd8e2f84ab27\") " pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:38:43 crc kubenswrapper[4806]: I1204 04:38:43.332839 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b94f0115-4cdc-4682-a897-cd8e2f84ab27-catalog-content\") pod \"redhat-operators-p52r9\" (UID: \"b94f0115-4cdc-4682-a897-cd8e2f84ab27\") " pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:38:43 crc kubenswrapper[4806]: I1204 04:38:43.353499 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlssj\" (UniqueName: \"kubernetes.io/projected/b94f0115-4cdc-4682-a897-cd8e2f84ab27-kube-api-access-rlssj\") pod \"redhat-operators-p52r9\" (UID: \"b94f0115-4cdc-4682-a897-cd8e2f84ab27\") " pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:38:43 crc kubenswrapper[4806]: I1204 04:38:43.517867 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:38:44 crc kubenswrapper[4806]: I1204 04:38:44.003839 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p52r9"] Dec 04 04:38:44 crc kubenswrapper[4806]: I1204 04:38:44.416736 4806 generic.go:334] "Generic (PLEG): container finished" podID="b94f0115-4cdc-4682-a897-cd8e2f84ab27" containerID="ddf4ca1c5854faab4f56485ffd2326e8ae3c2ee298960dd014890b2f511c7a0a" exitCode=0 Dec 04 04:38:44 crc kubenswrapper[4806]: I1204 04:38:44.416912 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p52r9" event={"ID":"b94f0115-4cdc-4682-a897-cd8e2f84ab27","Type":"ContainerDied","Data":"ddf4ca1c5854faab4f56485ffd2326e8ae3c2ee298960dd014890b2f511c7a0a"} Dec 04 04:38:44 crc kubenswrapper[4806]: I1204 04:38:44.417072 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p52r9" event={"ID":"b94f0115-4cdc-4682-a897-cd8e2f84ab27","Type":"ContainerStarted","Data":"393b8bc630fc7f685316ebd553dd1585e36f462dab8e505090afcc2f332f1798"} Dec 04 04:38:46 crc kubenswrapper[4806]: I1204 04:38:46.435217 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p52r9" event={"ID":"b94f0115-4cdc-4682-a897-cd8e2f84ab27","Type":"ContainerStarted","Data":"5919a56759d824618baf10a149ff46dbe2f03eb94d9f6666027404564ac052a7"} Dec 04 04:38:49 crc kubenswrapper[4806]: I1204 04:38:49.462232 4806 generic.go:334] "Generic (PLEG): container finished" podID="b94f0115-4cdc-4682-a897-cd8e2f84ab27" containerID="5919a56759d824618baf10a149ff46dbe2f03eb94d9f6666027404564ac052a7" exitCode=0 Dec 04 04:38:49 crc kubenswrapper[4806]: I1204 04:38:49.462314 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p52r9" event={"ID":"b94f0115-4cdc-4682-a897-cd8e2f84ab27","Type":"ContainerDied","Data":"5919a56759d824618baf10a149ff46dbe2f03eb94d9f6666027404564ac052a7"} Dec 04 04:38:49 crc kubenswrapper[4806]: I1204 04:38:49.466749 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 04:38:50 crc kubenswrapper[4806]: I1204 04:38:50.472606 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p52r9" event={"ID":"b94f0115-4cdc-4682-a897-cd8e2f84ab27","Type":"ContainerStarted","Data":"e08af53e7812332ba4dbf2d350d7e2d005ba4525096c0e043fb5bcbb24a4d1c8"} Dec 04 04:38:50 crc kubenswrapper[4806]: I1204 04:38:50.493247 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p52r9" podStartSLOduration=1.927348096 podStartE2EDuration="7.493231074s" podCreationTimestamp="2025-12-04 04:38:43 +0000 UTC" firstStartedPulling="2025-12-04 04:38:44.420426941 +0000 UTC m=+2639.278939859" lastFinishedPulling="2025-12-04 04:38:49.986309909 +0000 UTC m=+2644.844822837" observedRunningTime="2025-12-04 04:38:50.488093452 +0000 UTC m=+2645.346606390" watchObservedRunningTime="2025-12-04 04:38:50.493231074 +0000 UTC m=+2645.351744002" Dec 04 04:38:53 crc kubenswrapper[4806]: I1204 04:38:53.518495 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:38:53 crc kubenswrapper[4806]: I1204 04:38:53.520008 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:38:54 crc kubenswrapper[4806]: I1204 04:38:54.570111 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p52r9" podUID="b94f0115-4cdc-4682-a897-cd8e2f84ab27" containerName="registry-server" probeResult="failure" output=< Dec 04 04:38:54 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Dec 04 04:38:54 crc kubenswrapper[4806]: > Dec 04 04:38:57 crc kubenswrapper[4806]: I1204 04:38:57.047784 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:38:57 crc kubenswrapper[4806]: I1204 04:38:57.048181 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:39:03 crc kubenswrapper[4806]: I1204 04:39:03.573529 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:39:03 crc kubenswrapper[4806]: I1204 04:39:03.627262 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:39:03 crc kubenswrapper[4806]: I1204 04:39:03.812500 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p52r9"] Dec 04 04:39:04 crc kubenswrapper[4806]: I1204 04:39:04.594153 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p52r9" podUID="b94f0115-4cdc-4682-a897-cd8e2f84ab27" containerName="registry-server" containerID="cri-o://e08af53e7812332ba4dbf2d350d7e2d005ba4525096c0e043fb5bcbb24a4d1c8" gracePeriod=2 Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.027446 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.170143 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlssj\" (UniqueName: \"kubernetes.io/projected/b94f0115-4cdc-4682-a897-cd8e2f84ab27-kube-api-access-rlssj\") pod \"b94f0115-4cdc-4682-a897-cd8e2f84ab27\" (UID: \"b94f0115-4cdc-4682-a897-cd8e2f84ab27\") " Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.170427 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b94f0115-4cdc-4682-a897-cd8e2f84ab27-catalog-content\") pod \"b94f0115-4cdc-4682-a897-cd8e2f84ab27\" (UID: \"b94f0115-4cdc-4682-a897-cd8e2f84ab27\") " Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.170470 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b94f0115-4cdc-4682-a897-cd8e2f84ab27-utilities\") pod \"b94f0115-4cdc-4682-a897-cd8e2f84ab27\" (UID: \"b94f0115-4cdc-4682-a897-cd8e2f84ab27\") " Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.171656 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b94f0115-4cdc-4682-a897-cd8e2f84ab27-utilities" (OuterVolumeSpecName: "utilities") pod "b94f0115-4cdc-4682-a897-cd8e2f84ab27" (UID: "b94f0115-4cdc-4682-a897-cd8e2f84ab27"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.176870 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b94f0115-4cdc-4682-a897-cd8e2f84ab27-kube-api-access-rlssj" (OuterVolumeSpecName: "kube-api-access-rlssj") pod "b94f0115-4cdc-4682-a897-cd8e2f84ab27" (UID: "b94f0115-4cdc-4682-a897-cd8e2f84ab27"). InnerVolumeSpecName "kube-api-access-rlssj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.272277 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlssj\" (UniqueName: \"kubernetes.io/projected/b94f0115-4cdc-4682-a897-cd8e2f84ab27-kube-api-access-rlssj\") on node \"crc\" DevicePath \"\"" Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.272309 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b94f0115-4cdc-4682-a897-cd8e2f84ab27-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.293228 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b94f0115-4cdc-4682-a897-cd8e2f84ab27-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b94f0115-4cdc-4682-a897-cd8e2f84ab27" (UID: "b94f0115-4cdc-4682-a897-cd8e2f84ab27"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.374201 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b94f0115-4cdc-4682-a897-cd8e2f84ab27-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.606130 4806 generic.go:334] "Generic (PLEG): container finished" podID="b94f0115-4cdc-4682-a897-cd8e2f84ab27" containerID="e08af53e7812332ba4dbf2d350d7e2d005ba4525096c0e043fb5bcbb24a4d1c8" exitCode=0 Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.606187 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p52r9" event={"ID":"b94f0115-4cdc-4682-a897-cd8e2f84ab27","Type":"ContainerDied","Data":"e08af53e7812332ba4dbf2d350d7e2d005ba4525096c0e043fb5bcbb24a4d1c8"} Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.606221 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p52r9" event={"ID":"b94f0115-4cdc-4682-a897-cd8e2f84ab27","Type":"ContainerDied","Data":"393b8bc630fc7f685316ebd553dd1585e36f462dab8e505090afcc2f332f1798"} Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.606244 4806 scope.go:117] "RemoveContainer" containerID="e08af53e7812332ba4dbf2d350d7e2d005ba4525096c0e043fb5bcbb24a4d1c8" Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.606422 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p52r9" Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.639026 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p52r9"] Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.648595 4806 scope.go:117] "RemoveContainer" containerID="5919a56759d824618baf10a149ff46dbe2f03eb94d9f6666027404564ac052a7" Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.648688 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p52r9"] Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.668167 4806 scope.go:117] "RemoveContainer" containerID="ddf4ca1c5854faab4f56485ffd2326e8ae3c2ee298960dd014890b2f511c7a0a" Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.722121 4806 scope.go:117] "RemoveContainer" containerID="e08af53e7812332ba4dbf2d350d7e2d005ba4525096c0e043fb5bcbb24a4d1c8" Dec 04 04:39:05 crc kubenswrapper[4806]: E1204 04:39:05.722524 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e08af53e7812332ba4dbf2d350d7e2d005ba4525096c0e043fb5bcbb24a4d1c8\": container with ID starting with e08af53e7812332ba4dbf2d350d7e2d005ba4525096c0e043fb5bcbb24a4d1c8 not found: ID does not exist" containerID="e08af53e7812332ba4dbf2d350d7e2d005ba4525096c0e043fb5bcbb24a4d1c8" Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.722567 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e08af53e7812332ba4dbf2d350d7e2d005ba4525096c0e043fb5bcbb24a4d1c8"} err="failed to get container status \"e08af53e7812332ba4dbf2d350d7e2d005ba4525096c0e043fb5bcbb24a4d1c8\": rpc error: code = NotFound desc = could not find container \"e08af53e7812332ba4dbf2d350d7e2d005ba4525096c0e043fb5bcbb24a4d1c8\": container with ID starting with e08af53e7812332ba4dbf2d350d7e2d005ba4525096c0e043fb5bcbb24a4d1c8 not found: ID does not exist" Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.722591 4806 scope.go:117] "RemoveContainer" containerID="5919a56759d824618baf10a149ff46dbe2f03eb94d9f6666027404564ac052a7" Dec 04 04:39:05 crc kubenswrapper[4806]: E1204 04:39:05.723011 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5919a56759d824618baf10a149ff46dbe2f03eb94d9f6666027404564ac052a7\": container with ID starting with 5919a56759d824618baf10a149ff46dbe2f03eb94d9f6666027404564ac052a7 not found: ID does not exist" containerID="5919a56759d824618baf10a149ff46dbe2f03eb94d9f6666027404564ac052a7" Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.723115 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5919a56759d824618baf10a149ff46dbe2f03eb94d9f6666027404564ac052a7"} err="failed to get container status \"5919a56759d824618baf10a149ff46dbe2f03eb94d9f6666027404564ac052a7\": rpc error: code = NotFound desc = could not find container \"5919a56759d824618baf10a149ff46dbe2f03eb94d9f6666027404564ac052a7\": container with ID starting with 5919a56759d824618baf10a149ff46dbe2f03eb94d9f6666027404564ac052a7 not found: ID does not exist" Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.723202 4806 scope.go:117] "RemoveContainer" containerID="ddf4ca1c5854faab4f56485ffd2326e8ae3c2ee298960dd014890b2f511c7a0a" Dec 04 04:39:05 crc kubenswrapper[4806]: E1204 04:39:05.723577 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddf4ca1c5854faab4f56485ffd2326e8ae3c2ee298960dd014890b2f511c7a0a\": container with ID starting with ddf4ca1c5854faab4f56485ffd2326e8ae3c2ee298960dd014890b2f511c7a0a not found: ID does not exist" containerID="ddf4ca1c5854faab4f56485ffd2326e8ae3c2ee298960dd014890b2f511c7a0a" Dec 04 04:39:05 crc kubenswrapper[4806]: I1204 04:39:05.723602 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddf4ca1c5854faab4f56485ffd2326e8ae3c2ee298960dd014890b2f511c7a0a"} err="failed to get container status \"ddf4ca1c5854faab4f56485ffd2326e8ae3c2ee298960dd014890b2f511c7a0a\": rpc error: code = NotFound desc = could not find container \"ddf4ca1c5854faab4f56485ffd2326e8ae3c2ee298960dd014890b2f511c7a0a\": container with ID starting with ddf4ca1c5854faab4f56485ffd2326e8ae3c2ee298960dd014890b2f511c7a0a not found: ID does not exist" Dec 04 04:39:07 crc kubenswrapper[4806]: I1204 04:39:07.437942 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b94f0115-4cdc-4682-a897-cd8e2f84ab27" path="/var/lib/kubelet/pods/b94f0115-4cdc-4682-a897-cd8e2f84ab27/volumes" Dec 04 04:39:27 crc kubenswrapper[4806]: I1204 04:39:27.046997 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:39:27 crc kubenswrapper[4806]: I1204 04:39:27.047487 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:39:27 crc kubenswrapper[4806]: I1204 04:39:27.047534 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 04:39:27 crc kubenswrapper[4806]: I1204 04:39:27.048243 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cdb2e69904c8b363a34e509cfc2460a62051fc0403a906037069cecc07724904"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 04:39:27 crc kubenswrapper[4806]: I1204 04:39:27.048312 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://cdb2e69904c8b363a34e509cfc2460a62051fc0403a906037069cecc07724904" gracePeriod=600 Dec 04 04:39:27 crc kubenswrapper[4806]: I1204 04:39:27.816344 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="cdb2e69904c8b363a34e509cfc2460a62051fc0403a906037069cecc07724904" exitCode=0 Dec 04 04:39:27 crc kubenswrapper[4806]: I1204 04:39:27.816411 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"cdb2e69904c8b363a34e509cfc2460a62051fc0403a906037069cecc07724904"} Dec 04 04:39:27 crc kubenswrapper[4806]: I1204 04:39:27.817029 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f"} Dec 04 04:39:27 crc kubenswrapper[4806]: I1204 04:39:27.817054 4806 scope.go:117] "RemoveContainer" containerID="cdfec11f52fff3aaef374db48d72f6eb733d5293acedf9d8d45b0ee0cfba1513" Dec 04 04:41:27 crc kubenswrapper[4806]: I1204 04:41:27.047874 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:41:27 crc kubenswrapper[4806]: I1204 04:41:27.048393 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:41:57 crc kubenswrapper[4806]: I1204 04:41:57.047075 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:41:57 crc kubenswrapper[4806]: I1204 04:41:57.047704 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:41:58 crc kubenswrapper[4806]: I1204 04:41:58.212870 4806 generic.go:334] "Generic (PLEG): container finished" podID="21f3ccbb-669c-4055-b85c-f0c69fe61a8e" containerID="15c5843b9480b1fe84846753d32010e84d4b0b6f181d0964e77cc7db82d79ea4" exitCode=0 Dec 04 04:41:58 crc kubenswrapper[4806]: I1204 04:41:58.212993 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" event={"ID":"21f3ccbb-669c-4055-b85c-f0c69fe61a8e","Type":"ContainerDied","Data":"15c5843b9480b1fe84846753d32010e84d4b0b6f181d0964e77cc7db82d79ea4"} Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.669687 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.735497 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-migration-ssh-key-1\") pod \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.773959 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "21f3ccbb-669c-4055-b85c-f0c69fe61a8e" (UID: "21f3ccbb-669c-4055-b85c-f0c69fe61a8e"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.837349 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-cell1-compute-config-1\") pod \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.837419 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-inventory\") pod \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.837439 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-migration-ssh-key-0\") pod \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.837466 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-cell1-compute-config-0\") pod \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.837549 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrbb8\" (UniqueName: \"kubernetes.io/projected/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-kube-api-access-wrbb8\") pod \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.837683 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-ssh-key\") pod \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.837708 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-combined-ca-bundle\") pod \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.837777 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-extra-config-0\") pod \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\" (UID: \"21f3ccbb-669c-4055-b85c-f0c69fe61a8e\") " Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.838250 4806 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.847221 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "21f3ccbb-669c-4055-b85c-f0c69fe61a8e" (UID: "21f3ccbb-669c-4055-b85c-f0c69fe61a8e"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.847362 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-kube-api-access-wrbb8" (OuterVolumeSpecName: "kube-api-access-wrbb8") pod "21f3ccbb-669c-4055-b85c-f0c69fe61a8e" (UID: "21f3ccbb-669c-4055-b85c-f0c69fe61a8e"). InnerVolumeSpecName "kube-api-access-wrbb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.870683 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-inventory" (OuterVolumeSpecName: "inventory") pod "21f3ccbb-669c-4055-b85c-f0c69fe61a8e" (UID: "21f3ccbb-669c-4055-b85c-f0c69fe61a8e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.872126 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "21f3ccbb-669c-4055-b85c-f0c69fe61a8e" (UID: "21f3ccbb-669c-4055-b85c-f0c69fe61a8e"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.872766 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "21f3ccbb-669c-4055-b85c-f0c69fe61a8e" (UID: "21f3ccbb-669c-4055-b85c-f0c69fe61a8e"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.879492 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "21f3ccbb-669c-4055-b85c-f0c69fe61a8e" (UID: "21f3ccbb-669c-4055-b85c-f0c69fe61a8e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.880109 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "21f3ccbb-669c-4055-b85c-f0c69fe61a8e" (UID: "21f3ccbb-669c-4055-b85c-f0c69fe61a8e"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.881767 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "21f3ccbb-669c-4055-b85c-f0c69fe61a8e" (UID: "21f3ccbb-669c-4055-b85c-f0c69fe61a8e"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.940252 4806 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.940293 4806 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.940306 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.940317 4806 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.940329 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrbb8\" (UniqueName: \"kubernetes.io/projected/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-kube-api-access-wrbb8\") on node \"crc\" DevicePath \"\"" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.940339 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.940349 4806 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:41:59 crc kubenswrapper[4806]: I1204 04:41:59.940364 4806 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/21f3ccbb-669c-4055-b85c-f0c69fe61a8e-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.231971 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" event={"ID":"21f3ccbb-669c-4055-b85c-f0c69fe61a8e","Type":"ContainerDied","Data":"50e58dce49a9263258e2ae88322de0295cce329cddbd46116472df428ad005fd"} Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.232018 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50e58dce49a9263258e2ae88322de0295cce329cddbd46116472df428ad005fd" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.232086 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l4hb8" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.340945 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5"] Dec 04 04:42:00 crc kubenswrapper[4806]: E1204 04:42:00.341423 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b94f0115-4cdc-4682-a897-cd8e2f84ab27" containerName="extract-utilities" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.341448 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="b94f0115-4cdc-4682-a897-cd8e2f84ab27" containerName="extract-utilities" Dec 04 04:42:00 crc kubenswrapper[4806]: E1204 04:42:00.341465 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b94f0115-4cdc-4682-a897-cd8e2f84ab27" containerName="registry-server" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.341474 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="b94f0115-4cdc-4682-a897-cd8e2f84ab27" containerName="registry-server" Dec 04 04:42:00 crc kubenswrapper[4806]: E1204 04:42:00.341501 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b94f0115-4cdc-4682-a897-cd8e2f84ab27" containerName="extract-content" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.341511 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="b94f0115-4cdc-4682-a897-cd8e2f84ab27" containerName="extract-content" Dec 04 04:42:00 crc kubenswrapper[4806]: E1204 04:42:00.341535 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21f3ccbb-669c-4055-b85c-f0c69fe61a8e" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.341543 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="21f3ccbb-669c-4055-b85c-f0c69fe61a8e" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.341801 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="21f3ccbb-669c-4055-b85c-f0c69fe61a8e" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.341824 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="b94f0115-4cdc-4682-a897-cd8e2f84ab27" containerName="registry-server" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.342698 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.344885 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.344940 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.344999 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.345091 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.345394 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.345458 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.345525 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxrzj\" (UniqueName: \"kubernetes.io/projected/c9247c4e-1c6a-4110-8d97-6829c6064d48-kube-api-access-zxrzj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.346499 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.346847 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.347110 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.347342 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.347457 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-q8mqk" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.371667 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5"] Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.447437 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.447540 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.447593 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.447634 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxrzj\" (UniqueName: \"kubernetes.io/projected/c9247c4e-1c6a-4110-8d97-6829c6064d48-kube-api-access-zxrzj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.447685 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.447702 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.447734 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.455377 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.455439 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.455433 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.455522 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.457568 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.459457 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.469824 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxrzj\" (UniqueName: \"kubernetes.io/projected/c9247c4e-1c6a-4110-8d97-6829c6064d48-kube-api-access-zxrzj\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-stpm5\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:00 crc kubenswrapper[4806]: I1204 04:42:00.659033 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:42:01 crc kubenswrapper[4806]: I1204 04:42:01.302466 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5"] Dec 04 04:42:02 crc kubenswrapper[4806]: I1204 04:42:02.254782 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" event={"ID":"c9247c4e-1c6a-4110-8d97-6829c6064d48","Type":"ContainerStarted","Data":"2692e47b655d36ec0b4c3171572e5aef06eb345800430fc761599d4d1eaf0ba6"} Dec 04 04:42:02 crc kubenswrapper[4806]: I1204 04:42:02.255323 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" event={"ID":"c9247c4e-1c6a-4110-8d97-6829c6064d48","Type":"ContainerStarted","Data":"98d0dcc9a9bcfbd6d6dfa011713ee68113062bd26275c2e95b990f5bc0bec26d"} Dec 04 04:42:02 crc kubenswrapper[4806]: I1204 04:42:02.287350 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" podStartSLOduration=2.12404584 podStartE2EDuration="2.287328439s" podCreationTimestamp="2025-12-04 04:42:00 +0000 UTC" firstStartedPulling="2025-12-04 04:42:01.314949238 +0000 UTC m=+2836.173462166" lastFinishedPulling="2025-12-04 04:42:01.478231837 +0000 UTC m=+2836.336744765" observedRunningTime="2025-12-04 04:42:02.271484822 +0000 UTC m=+2837.129997780" watchObservedRunningTime="2025-12-04 04:42:02.287328439 +0000 UTC m=+2837.145841377" Dec 04 04:42:27 crc kubenswrapper[4806]: I1204 04:42:27.047947 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:42:27 crc kubenswrapper[4806]: I1204 04:42:27.048726 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:42:27 crc kubenswrapper[4806]: I1204 04:42:27.049231 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 04:42:27 crc kubenswrapper[4806]: I1204 04:42:27.050023 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 04:42:27 crc kubenswrapper[4806]: I1204 04:42:27.050088 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" gracePeriod=600 Dec 04 04:42:27 crc kubenswrapper[4806]: E1204 04:42:27.174757 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:42:27 crc kubenswrapper[4806]: I1204 04:42:27.495198 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" exitCode=0 Dec 04 04:42:27 crc kubenswrapper[4806]: I1204 04:42:27.495240 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f"} Dec 04 04:42:27 crc kubenswrapper[4806]: I1204 04:42:27.495272 4806 scope.go:117] "RemoveContainer" containerID="cdb2e69904c8b363a34e509cfc2460a62051fc0403a906037069cecc07724904" Dec 04 04:42:27 crc kubenswrapper[4806]: I1204 04:42:27.495881 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:42:27 crc kubenswrapper[4806]: E1204 04:42:27.496238 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:42:41 crc kubenswrapper[4806]: I1204 04:42:41.642164 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:42:41 crc kubenswrapper[4806]: E1204 04:42:41.643152 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:42:52 crc kubenswrapper[4806]: I1204 04:42:52.424352 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:42:52 crc kubenswrapper[4806]: E1204 04:42:52.425260 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:43:03 crc kubenswrapper[4806]: I1204 04:43:03.424188 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:43:03 crc kubenswrapper[4806]: E1204 04:43:03.425020 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:43:06 crc kubenswrapper[4806]: I1204 04:43:06.601856 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mr57v"] Dec 04 04:43:06 crc kubenswrapper[4806]: I1204 04:43:06.604275 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:06 crc kubenswrapper[4806]: I1204 04:43:06.613365 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mr57v"] Dec 04 04:43:06 crc kubenswrapper[4806]: I1204 04:43:06.724114 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95707a3f-82b5-4694-a703-0639c52e2b97-catalog-content\") pod \"community-operators-mr57v\" (UID: \"95707a3f-82b5-4694-a703-0639c52e2b97\") " pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:06 crc kubenswrapper[4806]: I1204 04:43:06.724296 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95707a3f-82b5-4694-a703-0639c52e2b97-utilities\") pod \"community-operators-mr57v\" (UID: \"95707a3f-82b5-4694-a703-0639c52e2b97\") " pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:06 crc kubenswrapper[4806]: I1204 04:43:06.724389 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwjtm\" (UniqueName: \"kubernetes.io/projected/95707a3f-82b5-4694-a703-0639c52e2b97-kube-api-access-wwjtm\") pod \"community-operators-mr57v\" (UID: \"95707a3f-82b5-4694-a703-0639c52e2b97\") " pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:06 crc kubenswrapper[4806]: I1204 04:43:06.826245 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95707a3f-82b5-4694-a703-0639c52e2b97-utilities\") pod \"community-operators-mr57v\" (UID: \"95707a3f-82b5-4694-a703-0639c52e2b97\") " pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:06 crc kubenswrapper[4806]: I1204 04:43:06.826565 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwjtm\" (UniqueName: \"kubernetes.io/projected/95707a3f-82b5-4694-a703-0639c52e2b97-kube-api-access-wwjtm\") pod \"community-operators-mr57v\" (UID: \"95707a3f-82b5-4694-a703-0639c52e2b97\") " pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:06 crc kubenswrapper[4806]: I1204 04:43:06.826734 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95707a3f-82b5-4694-a703-0639c52e2b97-catalog-content\") pod \"community-operators-mr57v\" (UID: \"95707a3f-82b5-4694-a703-0639c52e2b97\") " pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:06 crc kubenswrapper[4806]: I1204 04:43:06.827030 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95707a3f-82b5-4694-a703-0639c52e2b97-utilities\") pod \"community-operators-mr57v\" (UID: \"95707a3f-82b5-4694-a703-0639c52e2b97\") " pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:06 crc kubenswrapper[4806]: I1204 04:43:06.827117 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95707a3f-82b5-4694-a703-0639c52e2b97-catalog-content\") pod \"community-operators-mr57v\" (UID: \"95707a3f-82b5-4694-a703-0639c52e2b97\") " pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:06 crc kubenswrapper[4806]: I1204 04:43:06.851537 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwjtm\" (UniqueName: \"kubernetes.io/projected/95707a3f-82b5-4694-a703-0639c52e2b97-kube-api-access-wwjtm\") pod \"community-operators-mr57v\" (UID: \"95707a3f-82b5-4694-a703-0639c52e2b97\") " pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:06 crc kubenswrapper[4806]: I1204 04:43:06.923123 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:07 crc kubenswrapper[4806]: W1204 04:43:07.586693 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95707a3f_82b5_4694_a703_0639c52e2b97.slice/crio-07dfadc338b4dcfb0352592a0b8a195850cf4acfe6636bddefc24bc44ea75fb8 WatchSource:0}: Error finding container 07dfadc338b4dcfb0352592a0b8a195850cf4acfe6636bddefc24bc44ea75fb8: Status 404 returned error can't find the container with id 07dfadc338b4dcfb0352592a0b8a195850cf4acfe6636bddefc24bc44ea75fb8 Dec 04 04:43:07 crc kubenswrapper[4806]: I1204 04:43:07.586904 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mr57v"] Dec 04 04:43:07 crc kubenswrapper[4806]: I1204 04:43:07.876689 4806 generic.go:334] "Generic (PLEG): container finished" podID="95707a3f-82b5-4694-a703-0639c52e2b97" containerID="62a3db178eec73ca89cac2072ffdcbb3cf2d3a01632c41aa53cf4b33af88ed8a" exitCode=0 Dec 04 04:43:07 crc kubenswrapper[4806]: I1204 04:43:07.876742 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mr57v" event={"ID":"95707a3f-82b5-4694-a703-0639c52e2b97","Type":"ContainerDied","Data":"62a3db178eec73ca89cac2072ffdcbb3cf2d3a01632c41aa53cf4b33af88ed8a"} Dec 04 04:43:07 crc kubenswrapper[4806]: I1204 04:43:07.876779 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mr57v" event={"ID":"95707a3f-82b5-4694-a703-0639c52e2b97","Type":"ContainerStarted","Data":"07dfadc338b4dcfb0352592a0b8a195850cf4acfe6636bddefc24bc44ea75fb8"} Dec 04 04:43:09 crc kubenswrapper[4806]: I1204 04:43:09.909889 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mr57v" event={"ID":"95707a3f-82b5-4694-a703-0639c52e2b97","Type":"ContainerStarted","Data":"12674d372d09b7157072f2f41b6e80b5d98c379b71bf90c3fb568b1a47e39ac3"} Dec 04 04:43:10 crc kubenswrapper[4806]: I1204 04:43:10.920292 4806 generic.go:334] "Generic (PLEG): container finished" podID="95707a3f-82b5-4694-a703-0639c52e2b97" containerID="12674d372d09b7157072f2f41b6e80b5d98c379b71bf90c3fb568b1a47e39ac3" exitCode=0 Dec 04 04:43:10 crc kubenswrapper[4806]: I1204 04:43:10.920356 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mr57v" event={"ID":"95707a3f-82b5-4694-a703-0639c52e2b97","Type":"ContainerDied","Data":"12674d372d09b7157072f2f41b6e80b5d98c379b71bf90c3fb568b1a47e39ac3"} Dec 04 04:43:11 crc kubenswrapper[4806]: I1204 04:43:11.930774 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mr57v" event={"ID":"95707a3f-82b5-4694-a703-0639c52e2b97","Type":"ContainerStarted","Data":"f81f9533371b92585a055a11dc74fec25a73f3e749fd231fd8290d00c4b7fa1b"} Dec 04 04:43:11 crc kubenswrapper[4806]: I1204 04:43:11.948993 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mr57v" podStartSLOduration=2.434127307 podStartE2EDuration="5.948973972s" podCreationTimestamp="2025-12-04 04:43:06 +0000 UTC" firstStartedPulling="2025-12-04 04:43:07.878818935 +0000 UTC m=+2902.737331883" lastFinishedPulling="2025-12-04 04:43:11.39366562 +0000 UTC m=+2906.252178548" observedRunningTime="2025-12-04 04:43:11.945550154 +0000 UTC m=+2906.804063082" watchObservedRunningTime="2025-12-04 04:43:11.948973972 +0000 UTC m=+2906.807486910" Dec 04 04:43:14 crc kubenswrapper[4806]: I1204 04:43:14.423562 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:43:14 crc kubenswrapper[4806]: E1204 04:43:14.424339 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:43:16 crc kubenswrapper[4806]: I1204 04:43:16.923438 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:16 crc kubenswrapper[4806]: I1204 04:43:16.923778 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:16 crc kubenswrapper[4806]: I1204 04:43:16.985518 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:17 crc kubenswrapper[4806]: I1204 04:43:17.045019 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:17 crc kubenswrapper[4806]: I1204 04:43:17.228787 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mr57v"] Dec 04 04:43:18 crc kubenswrapper[4806]: I1204 04:43:18.989684 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mr57v" podUID="95707a3f-82b5-4694-a703-0639c52e2b97" containerName="registry-server" containerID="cri-o://f81f9533371b92585a055a11dc74fec25a73f3e749fd231fd8290d00c4b7fa1b" gracePeriod=2 Dec 04 04:43:19 crc kubenswrapper[4806]: I1204 04:43:19.528715 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:19 crc kubenswrapper[4806]: I1204 04:43:19.689059 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwjtm\" (UniqueName: \"kubernetes.io/projected/95707a3f-82b5-4694-a703-0639c52e2b97-kube-api-access-wwjtm\") pod \"95707a3f-82b5-4694-a703-0639c52e2b97\" (UID: \"95707a3f-82b5-4694-a703-0639c52e2b97\") " Dec 04 04:43:19 crc kubenswrapper[4806]: I1204 04:43:19.689118 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95707a3f-82b5-4694-a703-0639c52e2b97-utilities\") pod \"95707a3f-82b5-4694-a703-0639c52e2b97\" (UID: \"95707a3f-82b5-4694-a703-0639c52e2b97\") " Dec 04 04:43:19 crc kubenswrapper[4806]: I1204 04:43:19.689217 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95707a3f-82b5-4694-a703-0639c52e2b97-catalog-content\") pod \"95707a3f-82b5-4694-a703-0639c52e2b97\" (UID: \"95707a3f-82b5-4694-a703-0639c52e2b97\") " Dec 04 04:43:19 crc kubenswrapper[4806]: I1204 04:43:19.689600 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95707a3f-82b5-4694-a703-0639c52e2b97-utilities" (OuterVolumeSpecName: "utilities") pod "95707a3f-82b5-4694-a703-0639c52e2b97" (UID: "95707a3f-82b5-4694-a703-0639c52e2b97"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:43:19 crc kubenswrapper[4806]: I1204 04:43:19.698649 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95707a3f-82b5-4694-a703-0639c52e2b97-kube-api-access-wwjtm" (OuterVolumeSpecName: "kube-api-access-wwjtm") pod "95707a3f-82b5-4694-a703-0639c52e2b97" (UID: "95707a3f-82b5-4694-a703-0639c52e2b97"). InnerVolumeSpecName "kube-api-access-wwjtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:43:19 crc kubenswrapper[4806]: I1204 04:43:19.743650 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95707a3f-82b5-4694-a703-0639c52e2b97-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95707a3f-82b5-4694-a703-0639c52e2b97" (UID: "95707a3f-82b5-4694-a703-0639c52e2b97"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:43:19 crc kubenswrapper[4806]: I1204 04:43:19.791419 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwjtm\" (UniqueName: \"kubernetes.io/projected/95707a3f-82b5-4694-a703-0639c52e2b97-kube-api-access-wwjtm\") on node \"crc\" DevicePath \"\"" Dec 04 04:43:19 crc kubenswrapper[4806]: I1204 04:43:19.791459 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95707a3f-82b5-4694-a703-0639c52e2b97-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:43:19 crc kubenswrapper[4806]: I1204 04:43:19.791479 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95707a3f-82b5-4694-a703-0639c52e2b97-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:43:20 crc kubenswrapper[4806]: I1204 04:43:20.004541 4806 generic.go:334] "Generic (PLEG): container finished" podID="95707a3f-82b5-4694-a703-0639c52e2b97" containerID="f81f9533371b92585a055a11dc74fec25a73f3e749fd231fd8290d00c4b7fa1b" exitCode=0 Dec 04 04:43:20 crc kubenswrapper[4806]: I1204 04:43:20.004915 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mr57v" event={"ID":"95707a3f-82b5-4694-a703-0639c52e2b97","Type":"ContainerDied","Data":"f81f9533371b92585a055a11dc74fec25a73f3e749fd231fd8290d00c4b7fa1b"} Dec 04 04:43:20 crc kubenswrapper[4806]: I1204 04:43:20.005048 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mr57v" event={"ID":"95707a3f-82b5-4694-a703-0639c52e2b97","Type":"ContainerDied","Data":"07dfadc338b4dcfb0352592a0b8a195850cf4acfe6636bddefc24bc44ea75fb8"} Dec 04 04:43:20 crc kubenswrapper[4806]: I1204 04:43:20.005072 4806 scope.go:117] "RemoveContainer" containerID="f81f9533371b92585a055a11dc74fec25a73f3e749fd231fd8290d00c4b7fa1b" Dec 04 04:43:20 crc kubenswrapper[4806]: I1204 04:43:20.005262 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mr57v" Dec 04 04:43:20 crc kubenswrapper[4806]: I1204 04:43:20.036377 4806 scope.go:117] "RemoveContainer" containerID="12674d372d09b7157072f2f41b6e80b5d98c379b71bf90c3fb568b1a47e39ac3" Dec 04 04:43:20 crc kubenswrapper[4806]: I1204 04:43:20.039553 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mr57v"] Dec 04 04:43:20 crc kubenswrapper[4806]: I1204 04:43:20.049005 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mr57v"] Dec 04 04:43:20 crc kubenswrapper[4806]: I1204 04:43:20.061739 4806 scope.go:117] "RemoveContainer" containerID="62a3db178eec73ca89cac2072ffdcbb3cf2d3a01632c41aa53cf4b33af88ed8a" Dec 04 04:43:20 crc kubenswrapper[4806]: I1204 04:43:20.117471 4806 scope.go:117] "RemoveContainer" containerID="f81f9533371b92585a055a11dc74fec25a73f3e749fd231fd8290d00c4b7fa1b" Dec 04 04:43:20 crc kubenswrapper[4806]: E1204 04:43:20.117897 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f81f9533371b92585a055a11dc74fec25a73f3e749fd231fd8290d00c4b7fa1b\": container with ID starting with f81f9533371b92585a055a11dc74fec25a73f3e749fd231fd8290d00c4b7fa1b not found: ID does not exist" containerID="f81f9533371b92585a055a11dc74fec25a73f3e749fd231fd8290d00c4b7fa1b" Dec 04 04:43:20 crc kubenswrapper[4806]: I1204 04:43:20.117960 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f81f9533371b92585a055a11dc74fec25a73f3e749fd231fd8290d00c4b7fa1b"} err="failed to get container status \"f81f9533371b92585a055a11dc74fec25a73f3e749fd231fd8290d00c4b7fa1b\": rpc error: code = NotFound desc = could not find container \"f81f9533371b92585a055a11dc74fec25a73f3e749fd231fd8290d00c4b7fa1b\": container with ID starting with f81f9533371b92585a055a11dc74fec25a73f3e749fd231fd8290d00c4b7fa1b not found: ID does not exist" Dec 04 04:43:20 crc kubenswrapper[4806]: I1204 04:43:20.117990 4806 scope.go:117] "RemoveContainer" containerID="12674d372d09b7157072f2f41b6e80b5d98c379b71bf90c3fb568b1a47e39ac3" Dec 04 04:43:20 crc kubenswrapper[4806]: E1204 04:43:20.118315 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12674d372d09b7157072f2f41b6e80b5d98c379b71bf90c3fb568b1a47e39ac3\": container with ID starting with 12674d372d09b7157072f2f41b6e80b5d98c379b71bf90c3fb568b1a47e39ac3 not found: ID does not exist" containerID="12674d372d09b7157072f2f41b6e80b5d98c379b71bf90c3fb568b1a47e39ac3" Dec 04 04:43:20 crc kubenswrapper[4806]: I1204 04:43:20.118372 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12674d372d09b7157072f2f41b6e80b5d98c379b71bf90c3fb568b1a47e39ac3"} err="failed to get container status \"12674d372d09b7157072f2f41b6e80b5d98c379b71bf90c3fb568b1a47e39ac3\": rpc error: code = NotFound desc = could not find container \"12674d372d09b7157072f2f41b6e80b5d98c379b71bf90c3fb568b1a47e39ac3\": container with ID starting with 12674d372d09b7157072f2f41b6e80b5d98c379b71bf90c3fb568b1a47e39ac3 not found: ID does not exist" Dec 04 04:43:20 crc kubenswrapper[4806]: I1204 04:43:20.118402 4806 scope.go:117] "RemoveContainer" containerID="62a3db178eec73ca89cac2072ffdcbb3cf2d3a01632c41aa53cf4b33af88ed8a" Dec 04 04:43:20 crc kubenswrapper[4806]: E1204 04:43:20.118745 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62a3db178eec73ca89cac2072ffdcbb3cf2d3a01632c41aa53cf4b33af88ed8a\": container with ID starting with 62a3db178eec73ca89cac2072ffdcbb3cf2d3a01632c41aa53cf4b33af88ed8a not found: ID does not exist" containerID="62a3db178eec73ca89cac2072ffdcbb3cf2d3a01632c41aa53cf4b33af88ed8a" Dec 04 04:43:20 crc kubenswrapper[4806]: I1204 04:43:20.118870 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62a3db178eec73ca89cac2072ffdcbb3cf2d3a01632c41aa53cf4b33af88ed8a"} err="failed to get container status \"62a3db178eec73ca89cac2072ffdcbb3cf2d3a01632c41aa53cf4b33af88ed8a\": rpc error: code = NotFound desc = could not find container \"62a3db178eec73ca89cac2072ffdcbb3cf2d3a01632c41aa53cf4b33af88ed8a\": container with ID starting with 62a3db178eec73ca89cac2072ffdcbb3cf2d3a01632c41aa53cf4b33af88ed8a not found: ID does not exist" Dec 04 04:43:21 crc kubenswrapper[4806]: I1204 04:43:21.434514 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95707a3f-82b5-4694-a703-0639c52e2b97" path="/var/lib/kubelet/pods/95707a3f-82b5-4694-a703-0639c52e2b97/volumes" Dec 04 04:43:25 crc kubenswrapper[4806]: I1204 04:43:25.821919 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jw55m"] Dec 04 04:43:25 crc kubenswrapper[4806]: E1204 04:43:25.823012 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95707a3f-82b5-4694-a703-0639c52e2b97" containerName="extract-utilities" Dec 04 04:43:25 crc kubenswrapper[4806]: I1204 04:43:25.823030 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="95707a3f-82b5-4694-a703-0639c52e2b97" containerName="extract-utilities" Dec 04 04:43:25 crc kubenswrapper[4806]: E1204 04:43:25.823062 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95707a3f-82b5-4694-a703-0639c52e2b97" containerName="registry-server" Dec 04 04:43:25 crc kubenswrapper[4806]: I1204 04:43:25.823070 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="95707a3f-82b5-4694-a703-0639c52e2b97" containerName="registry-server" Dec 04 04:43:25 crc kubenswrapper[4806]: E1204 04:43:25.823093 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95707a3f-82b5-4694-a703-0639c52e2b97" containerName="extract-content" Dec 04 04:43:25 crc kubenswrapper[4806]: I1204 04:43:25.823104 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="95707a3f-82b5-4694-a703-0639c52e2b97" containerName="extract-content" Dec 04 04:43:25 crc kubenswrapper[4806]: I1204 04:43:25.823347 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="95707a3f-82b5-4694-a703-0639c52e2b97" containerName="registry-server" Dec 04 04:43:25 crc kubenswrapper[4806]: I1204 04:43:25.827698 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:25 crc kubenswrapper[4806]: I1204 04:43:25.836239 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jw55m"] Dec 04 04:43:26 crc kubenswrapper[4806]: I1204 04:43:26.010377 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-utilities\") pod \"redhat-marketplace-jw55m\" (UID: \"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90\") " pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:26 crc kubenswrapper[4806]: I1204 04:43:26.010963 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-catalog-content\") pod \"redhat-marketplace-jw55m\" (UID: \"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90\") " pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:26 crc kubenswrapper[4806]: I1204 04:43:26.011191 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nt72\" (UniqueName: \"kubernetes.io/projected/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-kube-api-access-2nt72\") pod \"redhat-marketplace-jw55m\" (UID: \"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90\") " pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:26 crc kubenswrapper[4806]: I1204 04:43:26.112378 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-catalog-content\") pod \"redhat-marketplace-jw55m\" (UID: \"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90\") " pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:26 crc kubenswrapper[4806]: I1204 04:43:26.112476 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nt72\" (UniqueName: \"kubernetes.io/projected/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-kube-api-access-2nt72\") pod \"redhat-marketplace-jw55m\" (UID: \"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90\") " pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:26 crc kubenswrapper[4806]: I1204 04:43:26.112576 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-utilities\") pod \"redhat-marketplace-jw55m\" (UID: \"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90\") " pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:26 crc kubenswrapper[4806]: I1204 04:43:26.112807 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-catalog-content\") pod \"redhat-marketplace-jw55m\" (UID: \"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90\") " pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:26 crc kubenswrapper[4806]: I1204 04:43:26.112942 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-utilities\") pod \"redhat-marketplace-jw55m\" (UID: \"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90\") " pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:26 crc kubenswrapper[4806]: I1204 04:43:26.134005 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nt72\" (UniqueName: \"kubernetes.io/projected/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-kube-api-access-2nt72\") pod \"redhat-marketplace-jw55m\" (UID: \"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90\") " pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:26 crc kubenswrapper[4806]: I1204 04:43:26.149979 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:26 crc kubenswrapper[4806]: I1204 04:43:26.629165 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jw55m"] Dec 04 04:43:27 crc kubenswrapper[4806]: I1204 04:43:27.069571 4806 generic.go:334] "Generic (PLEG): container finished" podID="c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90" containerID="48bfdc12db68a64e35d25763000d0dfd08c137774952e83088de5b977bd018b1" exitCode=0 Dec 04 04:43:27 crc kubenswrapper[4806]: I1204 04:43:27.069624 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jw55m" event={"ID":"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90","Type":"ContainerDied","Data":"48bfdc12db68a64e35d25763000d0dfd08c137774952e83088de5b977bd018b1"} Dec 04 04:43:27 crc kubenswrapper[4806]: I1204 04:43:27.069908 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jw55m" event={"ID":"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90","Type":"ContainerStarted","Data":"3634ca586a1f8ae8f9a77a4efd6dea26fbfa87cf48bdf374ce518c45bf196bdc"} Dec 04 04:43:27 crc kubenswrapper[4806]: I1204 04:43:27.424201 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:43:27 crc kubenswrapper[4806]: E1204 04:43:27.424743 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:43:28 crc kubenswrapper[4806]: I1204 04:43:28.083293 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jw55m" event={"ID":"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90","Type":"ContainerStarted","Data":"3d09a091e3d531dae4c10812ceb61a21d9c1667a05db19c8794519719fcc94a1"} Dec 04 04:43:29 crc kubenswrapper[4806]: I1204 04:43:29.097386 4806 generic.go:334] "Generic (PLEG): container finished" podID="c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90" containerID="3d09a091e3d531dae4c10812ceb61a21d9c1667a05db19c8794519719fcc94a1" exitCode=0 Dec 04 04:43:29 crc kubenswrapper[4806]: I1204 04:43:29.097486 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jw55m" event={"ID":"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90","Type":"ContainerDied","Data":"3d09a091e3d531dae4c10812ceb61a21d9c1667a05db19c8794519719fcc94a1"} Dec 04 04:43:30 crc kubenswrapper[4806]: I1204 04:43:30.108406 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jw55m" event={"ID":"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90","Type":"ContainerStarted","Data":"2d861e65f0a2f8e4c7e024dff9fe799ad3733e2de83464a0fbd7aa029512950d"} Dec 04 04:43:30 crc kubenswrapper[4806]: I1204 04:43:30.128543 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jw55m" podStartSLOduration=2.722611345 podStartE2EDuration="5.128525969s" podCreationTimestamp="2025-12-04 04:43:25 +0000 UTC" firstStartedPulling="2025-12-04 04:43:27.072172764 +0000 UTC m=+2921.930685692" lastFinishedPulling="2025-12-04 04:43:29.478087378 +0000 UTC m=+2924.336600316" observedRunningTime="2025-12-04 04:43:30.122954683 +0000 UTC m=+2924.981467631" watchObservedRunningTime="2025-12-04 04:43:30.128525969 +0000 UTC m=+2924.987038897" Dec 04 04:43:36 crc kubenswrapper[4806]: I1204 04:43:36.150978 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:36 crc kubenswrapper[4806]: I1204 04:43:36.151550 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:36 crc kubenswrapper[4806]: I1204 04:43:36.210134 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:36 crc kubenswrapper[4806]: I1204 04:43:36.260751 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:36 crc kubenswrapper[4806]: I1204 04:43:36.447615 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jw55m"] Dec 04 04:43:38 crc kubenswrapper[4806]: I1204 04:43:38.183973 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jw55m" podUID="c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90" containerName="registry-server" containerID="cri-o://2d861e65f0a2f8e4c7e024dff9fe799ad3733e2de83464a0fbd7aa029512950d" gracePeriod=2 Dec 04 04:43:38 crc kubenswrapper[4806]: I1204 04:43:38.654344 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:38 crc kubenswrapper[4806]: I1204 04:43:38.755236 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-catalog-content\") pod \"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90\" (UID: \"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90\") " Dec 04 04:43:38 crc kubenswrapper[4806]: I1204 04:43:38.755725 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nt72\" (UniqueName: \"kubernetes.io/projected/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-kube-api-access-2nt72\") pod \"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90\" (UID: \"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90\") " Dec 04 04:43:38 crc kubenswrapper[4806]: I1204 04:43:38.755857 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-utilities\") pod \"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90\" (UID: \"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90\") " Dec 04 04:43:38 crc kubenswrapper[4806]: I1204 04:43:38.756511 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-utilities" (OuterVolumeSpecName: "utilities") pod "c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90" (UID: "c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:43:38 crc kubenswrapper[4806]: I1204 04:43:38.763182 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-kube-api-access-2nt72" (OuterVolumeSpecName: "kube-api-access-2nt72") pod "c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90" (UID: "c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90"). InnerVolumeSpecName "kube-api-access-2nt72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:43:38 crc kubenswrapper[4806]: I1204 04:43:38.780458 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90" (UID: "c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:43:38 crc kubenswrapper[4806]: I1204 04:43:38.859008 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nt72\" (UniqueName: \"kubernetes.io/projected/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-kube-api-access-2nt72\") on node \"crc\" DevicePath \"\"" Dec 04 04:43:38 crc kubenswrapper[4806]: I1204 04:43:38.859067 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:43:38 crc kubenswrapper[4806]: I1204 04:43:38.859085 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.247658 4806 generic.go:334] "Generic (PLEG): container finished" podID="c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90" containerID="2d861e65f0a2f8e4c7e024dff9fe799ad3733e2de83464a0fbd7aa029512950d" exitCode=0 Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.247752 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jw55m" Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.247773 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jw55m" event={"ID":"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90","Type":"ContainerDied","Data":"2d861e65f0a2f8e4c7e024dff9fe799ad3733e2de83464a0fbd7aa029512950d"} Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.248682 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jw55m" event={"ID":"c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90","Type":"ContainerDied","Data":"3634ca586a1f8ae8f9a77a4efd6dea26fbfa87cf48bdf374ce518c45bf196bdc"} Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.248727 4806 scope.go:117] "RemoveContainer" containerID="2d861e65f0a2f8e4c7e024dff9fe799ad3733e2de83464a0fbd7aa029512950d" Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.298480 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jw55m"] Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.301080 4806 scope.go:117] "RemoveContainer" containerID="3d09a091e3d531dae4c10812ceb61a21d9c1667a05db19c8794519719fcc94a1" Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.305867 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jw55m"] Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.322004 4806 scope.go:117] "RemoveContainer" containerID="48bfdc12db68a64e35d25763000d0dfd08c137774952e83088de5b977bd018b1" Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.368205 4806 scope.go:117] "RemoveContainer" containerID="2d861e65f0a2f8e4c7e024dff9fe799ad3733e2de83464a0fbd7aa029512950d" Dec 04 04:43:39 crc kubenswrapper[4806]: E1204 04:43:39.369002 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d861e65f0a2f8e4c7e024dff9fe799ad3733e2de83464a0fbd7aa029512950d\": container with ID starting with 2d861e65f0a2f8e4c7e024dff9fe799ad3733e2de83464a0fbd7aa029512950d not found: ID does not exist" containerID="2d861e65f0a2f8e4c7e024dff9fe799ad3733e2de83464a0fbd7aa029512950d" Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.369049 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d861e65f0a2f8e4c7e024dff9fe799ad3733e2de83464a0fbd7aa029512950d"} err="failed to get container status \"2d861e65f0a2f8e4c7e024dff9fe799ad3733e2de83464a0fbd7aa029512950d\": rpc error: code = NotFound desc = could not find container \"2d861e65f0a2f8e4c7e024dff9fe799ad3733e2de83464a0fbd7aa029512950d\": container with ID starting with 2d861e65f0a2f8e4c7e024dff9fe799ad3733e2de83464a0fbd7aa029512950d not found: ID does not exist" Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.369075 4806 scope.go:117] "RemoveContainer" containerID="3d09a091e3d531dae4c10812ceb61a21d9c1667a05db19c8794519719fcc94a1" Dec 04 04:43:39 crc kubenswrapper[4806]: E1204 04:43:39.369349 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d09a091e3d531dae4c10812ceb61a21d9c1667a05db19c8794519719fcc94a1\": container with ID starting with 3d09a091e3d531dae4c10812ceb61a21d9c1667a05db19c8794519719fcc94a1 not found: ID does not exist" containerID="3d09a091e3d531dae4c10812ceb61a21d9c1667a05db19c8794519719fcc94a1" Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.369393 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d09a091e3d531dae4c10812ceb61a21d9c1667a05db19c8794519719fcc94a1"} err="failed to get container status \"3d09a091e3d531dae4c10812ceb61a21d9c1667a05db19c8794519719fcc94a1\": rpc error: code = NotFound desc = could not find container \"3d09a091e3d531dae4c10812ceb61a21d9c1667a05db19c8794519719fcc94a1\": container with ID starting with 3d09a091e3d531dae4c10812ceb61a21d9c1667a05db19c8794519719fcc94a1 not found: ID does not exist" Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.369424 4806 scope.go:117] "RemoveContainer" containerID="48bfdc12db68a64e35d25763000d0dfd08c137774952e83088de5b977bd018b1" Dec 04 04:43:39 crc kubenswrapper[4806]: E1204 04:43:39.369704 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48bfdc12db68a64e35d25763000d0dfd08c137774952e83088de5b977bd018b1\": container with ID starting with 48bfdc12db68a64e35d25763000d0dfd08c137774952e83088de5b977bd018b1 not found: ID does not exist" containerID="48bfdc12db68a64e35d25763000d0dfd08c137774952e83088de5b977bd018b1" Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.369771 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48bfdc12db68a64e35d25763000d0dfd08c137774952e83088de5b977bd018b1"} err="failed to get container status \"48bfdc12db68a64e35d25763000d0dfd08c137774952e83088de5b977bd018b1\": rpc error: code = NotFound desc = could not find container \"48bfdc12db68a64e35d25763000d0dfd08c137774952e83088de5b977bd018b1\": container with ID starting with 48bfdc12db68a64e35d25763000d0dfd08c137774952e83088de5b977bd018b1 not found: ID does not exist" Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.424601 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:43:39 crc kubenswrapper[4806]: E1204 04:43:39.425065 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:43:39 crc kubenswrapper[4806]: I1204 04:43:39.435150 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90" path="/var/lib/kubelet/pods/c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90/volumes" Dec 04 04:43:53 crc kubenswrapper[4806]: I1204 04:43:53.424368 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:43:53 crc kubenswrapper[4806]: E1204 04:43:53.425245 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:44:05 crc kubenswrapper[4806]: I1204 04:44:05.424066 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:44:05 crc kubenswrapper[4806]: E1204 04:44:05.424825 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:44:16 crc kubenswrapper[4806]: I1204 04:44:16.424587 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:44:16 crc kubenswrapper[4806]: E1204 04:44:16.426107 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:44:30 crc kubenswrapper[4806]: I1204 04:44:30.424428 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:44:30 crc kubenswrapper[4806]: E1204 04:44:30.426641 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:44:43 crc kubenswrapper[4806]: I1204 04:44:43.424723 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:44:43 crc kubenswrapper[4806]: E1204 04:44:43.425599 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:44:58 crc kubenswrapper[4806]: I1204 04:44:58.425255 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:44:58 crc kubenswrapper[4806]: E1204 04:44:58.426466 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.152377 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll"] Dec 04 04:45:00 crc kubenswrapper[4806]: E1204 04:45:00.154209 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90" containerName="extract-utilities" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.154311 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90" containerName="extract-utilities" Dec 04 04:45:00 crc kubenswrapper[4806]: E1204 04:45:00.154390 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90" containerName="registry-server" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.154467 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90" containerName="registry-server" Dec 04 04:45:00 crc kubenswrapper[4806]: E1204 04:45:00.154535 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90" containerName="extract-content" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.154590 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90" containerName="extract-content" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.154862 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="c85c18eb-7e26-4adf-aa6a-c3bb9d63fb90" containerName="registry-server" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.155601 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.158486 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.158803 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.174029 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll"] Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.250833 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-config-volume\") pod \"collect-profiles-29413725-4vnll\" (UID: \"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.251480 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-secret-volume\") pod \"collect-profiles-29413725-4vnll\" (UID: \"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.352961 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-secret-volume\") pod \"collect-profiles-29413725-4vnll\" (UID: \"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.353145 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdght\" (UniqueName: \"kubernetes.io/projected/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-kube-api-access-hdght\") pod \"collect-profiles-29413725-4vnll\" (UID: \"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.353178 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-config-volume\") pod \"collect-profiles-29413725-4vnll\" (UID: \"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.354258 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-config-volume\") pod \"collect-profiles-29413725-4vnll\" (UID: \"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.359610 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-secret-volume\") pod \"collect-profiles-29413725-4vnll\" (UID: \"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.454626 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdght\" (UniqueName: \"kubernetes.io/projected/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-kube-api-access-hdght\") pod \"collect-profiles-29413725-4vnll\" (UID: \"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.475737 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdght\" (UniqueName: \"kubernetes.io/projected/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-kube-api-access-hdght\") pod \"collect-profiles-29413725-4vnll\" (UID: \"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.482427 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll" Dec 04 04:45:00 crc kubenswrapper[4806]: I1204 04:45:00.965750 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll"] Dec 04 04:45:01 crc kubenswrapper[4806]: I1204 04:45:01.925034 4806 generic.go:334] "Generic (PLEG): container finished" podID="c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679" containerID="28094f5cafb39713eed10d3b7ccffa5f7e4a97fb80832b06fd4c4c92e2759490" exitCode=0 Dec 04 04:45:01 crc kubenswrapper[4806]: I1204 04:45:01.925134 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll" event={"ID":"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679","Type":"ContainerDied","Data":"28094f5cafb39713eed10d3b7ccffa5f7e4a97fb80832b06fd4c4c92e2759490"} Dec 04 04:45:01 crc kubenswrapper[4806]: I1204 04:45:01.925406 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll" event={"ID":"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679","Type":"ContainerStarted","Data":"42cfc1c25a87f4e83d76afbc11cfe8b08be6bc169a1603e29a3ba5aba7bff519"} Dec 04 04:45:03 crc kubenswrapper[4806]: I1204 04:45:03.286939 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll" Dec 04 04:45:03 crc kubenswrapper[4806]: I1204 04:45:03.419220 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-secret-volume\") pod \"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679\" (UID: \"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679\") " Dec 04 04:45:03 crc kubenswrapper[4806]: I1204 04:45:03.419313 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdght\" (UniqueName: \"kubernetes.io/projected/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-kube-api-access-hdght\") pod \"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679\" (UID: \"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679\") " Dec 04 04:45:03 crc kubenswrapper[4806]: I1204 04:45:03.419368 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-config-volume\") pod \"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679\" (UID: \"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679\") " Dec 04 04:45:03 crc kubenswrapper[4806]: I1204 04:45:03.420188 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-config-volume" (OuterVolumeSpecName: "config-volume") pod "c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679" (UID: "c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 04:45:03 crc kubenswrapper[4806]: I1204 04:45:03.428274 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679" (UID: "c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:45:03 crc kubenswrapper[4806]: I1204 04:45:03.441315 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-kube-api-access-hdght" (OuterVolumeSpecName: "kube-api-access-hdght") pod "c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679" (UID: "c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679"). InnerVolumeSpecName "kube-api-access-hdght". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:45:03 crc kubenswrapper[4806]: I1204 04:45:03.521721 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdght\" (UniqueName: \"kubernetes.io/projected/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-kube-api-access-hdght\") on node \"crc\" DevicePath \"\"" Dec 04 04:45:03 crc kubenswrapper[4806]: I1204 04:45:03.521765 4806 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 04:45:03 crc kubenswrapper[4806]: I1204 04:45:03.521781 4806 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 04:45:03 crc kubenswrapper[4806]: I1204 04:45:03.947016 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll" event={"ID":"c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679","Type":"ContainerDied","Data":"42cfc1c25a87f4e83d76afbc11cfe8b08be6bc169a1603e29a3ba5aba7bff519"} Dec 04 04:45:03 crc kubenswrapper[4806]: I1204 04:45:03.947053 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll" Dec 04 04:45:03 crc kubenswrapper[4806]: I1204 04:45:03.947058 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42cfc1c25a87f4e83d76afbc11cfe8b08be6bc169a1603e29a3ba5aba7bff519" Dec 04 04:45:04 crc kubenswrapper[4806]: I1204 04:45:04.367491 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw"] Dec 04 04:45:04 crc kubenswrapper[4806]: I1204 04:45:04.377634 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413680-2hjnw"] Dec 04 04:45:05 crc kubenswrapper[4806]: I1204 04:45:05.436444 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2da7dc1-86d1-4bbf-964d-9d9f83b17757" path="/var/lib/kubelet/pods/d2da7dc1-86d1-4bbf-964d-9d9f83b17757/volumes" Dec 04 04:45:09 crc kubenswrapper[4806]: I1204 04:45:09.424542 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:45:09 crc kubenswrapper[4806]: E1204 04:45:09.425551 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:45:24 crc kubenswrapper[4806]: I1204 04:45:24.424412 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:45:24 crc kubenswrapper[4806]: E1204 04:45:24.425455 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:45:35 crc kubenswrapper[4806]: I1204 04:45:35.432215 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:45:35 crc kubenswrapper[4806]: E1204 04:45:35.436450 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:45:41 crc kubenswrapper[4806]: I1204 04:45:41.957244 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-cj594" podUID="5e02c1ac-06b2-4d89-98f8-545f682c4a4d" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 04:45:47 crc kubenswrapper[4806]: I1204 04:45:47.424266 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:45:47 crc kubenswrapper[4806]: E1204 04:45:47.424867 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:45:58 crc kubenswrapper[4806]: I1204 04:45:58.423917 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:45:58 crc kubenswrapper[4806]: E1204 04:45:58.424983 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:46:01 crc kubenswrapper[4806]: I1204 04:46:01.905224 4806 scope.go:117] "RemoveContainer" containerID="34f58022124a4c35a0592a1f8fabe1e14b3fd4726b773fb9c61f91f509cfd15c" Dec 04 04:46:05 crc kubenswrapper[4806]: I1204 04:46:05.493610 4806 generic.go:334] "Generic (PLEG): container finished" podID="c9247c4e-1c6a-4110-8d97-6829c6064d48" containerID="2692e47b655d36ec0b4c3171572e5aef06eb345800430fc761599d4d1eaf0ba6" exitCode=0 Dec 04 04:46:05 crc kubenswrapper[4806]: I1204 04:46:05.493664 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" event={"ID":"c9247c4e-1c6a-4110-8d97-6829c6064d48","Type":"ContainerDied","Data":"2692e47b655d36ec0b4c3171572e5aef06eb345800430fc761599d4d1eaf0ba6"} Dec 04 04:46:06 crc kubenswrapper[4806]: I1204 04:46:06.925050 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.089469 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-1\") pod \"c9247c4e-1c6a-4110-8d97-6829c6064d48\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.089519 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-0\") pod \"c9247c4e-1c6a-4110-8d97-6829c6064d48\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.089570 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-2\") pod \"c9247c4e-1c6a-4110-8d97-6829c6064d48\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.089601 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-telemetry-combined-ca-bundle\") pod \"c9247c4e-1c6a-4110-8d97-6829c6064d48\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.089714 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ssh-key\") pod \"c9247c4e-1c6a-4110-8d97-6829c6064d48\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.089746 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxrzj\" (UniqueName: \"kubernetes.io/projected/c9247c4e-1c6a-4110-8d97-6829c6064d48-kube-api-access-zxrzj\") pod \"c9247c4e-1c6a-4110-8d97-6829c6064d48\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.089784 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-inventory\") pod \"c9247c4e-1c6a-4110-8d97-6829c6064d48\" (UID: \"c9247c4e-1c6a-4110-8d97-6829c6064d48\") " Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.097384 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9247c4e-1c6a-4110-8d97-6829c6064d48-kube-api-access-zxrzj" (OuterVolumeSpecName: "kube-api-access-zxrzj") pod "c9247c4e-1c6a-4110-8d97-6829c6064d48" (UID: "c9247c4e-1c6a-4110-8d97-6829c6064d48"). InnerVolumeSpecName "kube-api-access-zxrzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.097592 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c9247c4e-1c6a-4110-8d97-6829c6064d48" (UID: "c9247c4e-1c6a-4110-8d97-6829c6064d48"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.120692 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-inventory" (OuterVolumeSpecName: "inventory") pod "c9247c4e-1c6a-4110-8d97-6829c6064d48" (UID: "c9247c4e-1c6a-4110-8d97-6829c6064d48"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.121596 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "c9247c4e-1c6a-4110-8d97-6829c6064d48" (UID: "c9247c4e-1c6a-4110-8d97-6829c6064d48"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.124873 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c9247c4e-1c6a-4110-8d97-6829c6064d48" (UID: "c9247c4e-1c6a-4110-8d97-6829c6064d48"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.127878 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "c9247c4e-1c6a-4110-8d97-6829c6064d48" (UID: "c9247c4e-1c6a-4110-8d97-6829c6064d48"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.129458 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "c9247c4e-1c6a-4110-8d97-6829c6064d48" (UID: "c9247c4e-1c6a-4110-8d97-6829c6064d48"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.192307 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.192587 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxrzj\" (UniqueName: \"kubernetes.io/projected/c9247c4e-1c6a-4110-8d97-6829c6064d48-kube-api-access-zxrzj\") on node \"crc\" DevicePath \"\"" Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.192716 4806 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.192810 4806 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.192890 4806 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.193031 4806 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.193115 4806 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9247c4e-1c6a-4110-8d97-6829c6064d48-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.516585 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" event={"ID":"c9247c4e-1c6a-4110-8d97-6829c6064d48","Type":"ContainerDied","Data":"98d0dcc9a9bcfbd6d6dfa011713ee68113062bd26275c2e95b990f5bc0bec26d"} Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.516629 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98d0dcc9a9bcfbd6d6dfa011713ee68113062bd26275c2e95b990f5bc0bec26d" Dec 04 04:46:07 crc kubenswrapper[4806]: I1204 04:46:07.516639 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-stpm5" Dec 04 04:46:11 crc kubenswrapper[4806]: I1204 04:46:11.424706 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:46:11 crc kubenswrapper[4806]: E1204 04:46:11.425417 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:46:23 crc kubenswrapper[4806]: I1204 04:46:23.424296 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:46:23 crc kubenswrapper[4806]: E1204 04:46:23.427415 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:46:36 crc kubenswrapper[4806]: I1204 04:46:36.423728 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:46:36 crc kubenswrapper[4806]: E1204 04:46:36.424389 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:46:47 crc kubenswrapper[4806]: I1204 04:46:47.423610 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:46:47 crc kubenswrapper[4806]: E1204 04:46:47.424584 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:46:58 crc kubenswrapper[4806]: I1204 04:46:58.424465 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:46:58 crc kubenswrapper[4806]: E1204 04:46:58.426570 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.548674 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 04 04:47:04 crc kubenswrapper[4806]: E1204 04:47:04.549425 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9247c4e-1c6a-4110-8d97-6829c6064d48" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.549443 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9247c4e-1c6a-4110-8d97-6829c6064d48" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 04 04:47:04 crc kubenswrapper[4806]: E1204 04:47:04.549483 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679" containerName="collect-profiles" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.549491 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679" containerName="collect-profiles" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.549699 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9247c4e-1c6a-4110-8d97-6829c6064d48" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.549732 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679" containerName="collect-profiles" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.550588 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.597800 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.598027 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-jb6sp" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.598188 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.598330 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.610403 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.634904 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.638597 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/319a7357-70c6-4b54-8e97-9860dd0618e7-config-data\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.638874 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/319a7357-70c6-4b54-8e97-9860dd0618e7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.742030 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/319a7357-70c6-4b54-8e97-9860dd0618e7-config-data\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.742353 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/319a7357-70c6-4b54-8e97-9860dd0618e7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.742653 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.742697 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/319a7357-70c6-4b54-8e97-9860dd0618e7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.742713 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.742756 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.742807 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njg9d\" (UniqueName: \"kubernetes.io/projected/319a7357-70c6-4b54-8e97-9860dd0618e7-kube-api-access-njg9d\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.742824 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.742842 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/319a7357-70c6-4b54-8e97-9860dd0618e7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.743700 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/319a7357-70c6-4b54-8e97-9860dd0618e7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.744445 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/319a7357-70c6-4b54-8e97-9860dd0618e7-config-data\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.749233 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.844529 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njg9d\" (UniqueName: \"kubernetes.io/projected/319a7357-70c6-4b54-8e97-9860dd0618e7-kube-api-access-njg9d\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.844567 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.844587 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/319a7357-70c6-4b54-8e97-9860dd0618e7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.844670 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/319a7357-70c6-4b54-8e97-9860dd0618e7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.844714 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.844730 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.845116 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/319a7357-70c6-4b54-8e97-9860dd0618e7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.845381 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.845410 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/319a7357-70c6-4b54-8e97-9860dd0618e7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.853647 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.853897 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.873032 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njg9d\" (UniqueName: \"kubernetes.io/projected/319a7357-70c6-4b54-8e97-9860dd0618e7-kube-api-access-njg9d\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.873413 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " pod="openstack/tempest-tests-tempest" Dec 04 04:47:04 crc kubenswrapper[4806]: I1204 04:47:04.927259 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 04 04:47:05 crc kubenswrapper[4806]: I1204 04:47:05.359760 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 04 04:47:05 crc kubenswrapper[4806]: I1204 04:47:05.373279 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 04:47:06 crc kubenswrapper[4806]: I1204 04:47:06.053126 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"319a7357-70c6-4b54-8e97-9860dd0618e7","Type":"ContainerStarted","Data":"150f0528353b9920c10e0f675e8348ce6f78c30ca3ac89fc7ba1504ff7b89cf2"} Dec 04 04:47:10 crc kubenswrapper[4806]: I1204 04:47:10.424286 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:47:10 crc kubenswrapper[4806]: E1204 04:47:10.426438 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:47:23 crc kubenswrapper[4806]: I1204 04:47:23.423629 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:47:23 crc kubenswrapper[4806]: E1204 04:47:23.424454 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:47:38 crc kubenswrapper[4806]: I1204 04:47:38.423868 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:47:43 crc kubenswrapper[4806]: E1204 04:47:43.442832 4806 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 04 04:47:43 crc kubenswrapper[4806]: E1204 04:47:43.444585 4806 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-njg9d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(319a7357-70c6-4b54-8e97-9860dd0618e7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 04:47:43 crc kubenswrapper[4806]: E1204 04:47:43.446278 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="319a7357-70c6-4b54-8e97-9860dd0618e7" Dec 04 04:47:44 crc kubenswrapper[4806]: I1204 04:47:44.475205 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"180699dbf0c533c23b37a9f63f83cd78ab2f3830f234d8427f5b5bc67b2a3005"} Dec 04 04:47:44 crc kubenswrapper[4806]: E1204 04:47:44.478626 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="319a7357-70c6-4b54-8e97-9860dd0618e7" Dec 04 04:47:57 crc kubenswrapper[4806]: I1204 04:47:57.851040 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 04 04:47:59 crc kubenswrapper[4806]: I1204 04:47:59.603416 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"319a7357-70c6-4b54-8e97-9860dd0618e7","Type":"ContainerStarted","Data":"53bbe8fc4a8361c9961c95aec4994de66b7c8f7d18efb9bbe065be6e3ba94996"} Dec 04 04:47:59 crc kubenswrapper[4806]: I1204 04:47:59.642705 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.167541881 podStartE2EDuration="56.642683307s" podCreationTimestamp="2025-12-04 04:47:03 +0000 UTC" firstStartedPulling="2025-12-04 04:47:05.371813376 +0000 UTC m=+3140.230326304" lastFinishedPulling="2025-12-04 04:47:57.846954802 +0000 UTC m=+3192.705467730" observedRunningTime="2025-12-04 04:47:59.625701293 +0000 UTC m=+3194.484214221" watchObservedRunningTime="2025-12-04 04:47:59.642683307 +0000 UTC m=+3194.501196235" Dec 04 04:48:35 crc kubenswrapper[4806]: I1204 04:48:35.857071 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8ldlh"] Dec 04 04:48:35 crc kubenswrapper[4806]: I1204 04:48:35.860643 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:35 crc kubenswrapper[4806]: I1204 04:48:35.896420 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8ldlh"] Dec 04 04:48:36 crc kubenswrapper[4806]: I1204 04:48:36.028222 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw6fx\" (UniqueName: \"kubernetes.io/projected/2a187bb3-1801-491d-82a2-c0f1ac85ac83-kube-api-access-lw6fx\") pod \"certified-operators-8ldlh\" (UID: \"2a187bb3-1801-491d-82a2-c0f1ac85ac83\") " pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:36 crc kubenswrapper[4806]: I1204 04:48:36.028271 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a187bb3-1801-491d-82a2-c0f1ac85ac83-catalog-content\") pod \"certified-operators-8ldlh\" (UID: \"2a187bb3-1801-491d-82a2-c0f1ac85ac83\") " pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:36 crc kubenswrapper[4806]: I1204 04:48:36.028367 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a187bb3-1801-491d-82a2-c0f1ac85ac83-utilities\") pod \"certified-operators-8ldlh\" (UID: \"2a187bb3-1801-491d-82a2-c0f1ac85ac83\") " pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:36 crc kubenswrapper[4806]: I1204 04:48:36.130177 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a187bb3-1801-491d-82a2-c0f1ac85ac83-utilities\") pod \"certified-operators-8ldlh\" (UID: \"2a187bb3-1801-491d-82a2-c0f1ac85ac83\") " pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:36 crc kubenswrapper[4806]: I1204 04:48:36.130326 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw6fx\" (UniqueName: \"kubernetes.io/projected/2a187bb3-1801-491d-82a2-c0f1ac85ac83-kube-api-access-lw6fx\") pod \"certified-operators-8ldlh\" (UID: \"2a187bb3-1801-491d-82a2-c0f1ac85ac83\") " pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:36 crc kubenswrapper[4806]: I1204 04:48:36.130348 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a187bb3-1801-491d-82a2-c0f1ac85ac83-catalog-content\") pod \"certified-operators-8ldlh\" (UID: \"2a187bb3-1801-491d-82a2-c0f1ac85ac83\") " pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:36 crc kubenswrapper[4806]: I1204 04:48:36.130825 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a187bb3-1801-491d-82a2-c0f1ac85ac83-catalog-content\") pod \"certified-operators-8ldlh\" (UID: \"2a187bb3-1801-491d-82a2-c0f1ac85ac83\") " pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:36 crc kubenswrapper[4806]: I1204 04:48:36.131063 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a187bb3-1801-491d-82a2-c0f1ac85ac83-utilities\") pod \"certified-operators-8ldlh\" (UID: \"2a187bb3-1801-491d-82a2-c0f1ac85ac83\") " pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:36 crc kubenswrapper[4806]: I1204 04:48:36.154313 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw6fx\" (UniqueName: \"kubernetes.io/projected/2a187bb3-1801-491d-82a2-c0f1ac85ac83-kube-api-access-lw6fx\") pod \"certified-operators-8ldlh\" (UID: \"2a187bb3-1801-491d-82a2-c0f1ac85ac83\") " pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:36 crc kubenswrapper[4806]: I1204 04:48:36.180736 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:37 crc kubenswrapper[4806]: I1204 04:48:37.250635 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8ldlh"] Dec 04 04:48:37 crc kubenswrapper[4806]: I1204 04:48:37.940769 4806 generic.go:334] "Generic (PLEG): container finished" podID="2a187bb3-1801-491d-82a2-c0f1ac85ac83" containerID="be080212336c6ad4e59b26baff5d0fe23ac2bb77e541ace6a947e9e39916bf3d" exitCode=0 Dec 04 04:48:37 crc kubenswrapper[4806]: I1204 04:48:37.940819 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ldlh" event={"ID":"2a187bb3-1801-491d-82a2-c0f1ac85ac83","Type":"ContainerDied","Data":"be080212336c6ad4e59b26baff5d0fe23ac2bb77e541ace6a947e9e39916bf3d"} Dec 04 04:48:37 crc kubenswrapper[4806]: I1204 04:48:37.941508 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ldlh" event={"ID":"2a187bb3-1801-491d-82a2-c0f1ac85ac83","Type":"ContainerStarted","Data":"44ab2e04a73418783dd34b4509e767a275a26f371a45bbc3986e1d681ab4be43"} Dec 04 04:48:38 crc kubenswrapper[4806]: I1204 04:48:38.960950 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ldlh" event={"ID":"2a187bb3-1801-491d-82a2-c0f1ac85ac83","Type":"ContainerStarted","Data":"1e1c3b6fff95616d6c78a8dd542cce8a20af42c80f86d809a70a25023c999c85"} Dec 04 04:48:40 crc kubenswrapper[4806]: I1204 04:48:40.982893 4806 generic.go:334] "Generic (PLEG): container finished" podID="2a187bb3-1801-491d-82a2-c0f1ac85ac83" containerID="1e1c3b6fff95616d6c78a8dd542cce8a20af42c80f86d809a70a25023c999c85" exitCode=0 Dec 04 04:48:40 crc kubenswrapper[4806]: I1204 04:48:40.982996 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ldlh" event={"ID":"2a187bb3-1801-491d-82a2-c0f1ac85ac83","Type":"ContainerDied","Data":"1e1c3b6fff95616d6c78a8dd542cce8a20af42c80f86d809a70a25023c999c85"} Dec 04 04:48:41 crc kubenswrapper[4806]: I1204 04:48:41.999242 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ldlh" event={"ID":"2a187bb3-1801-491d-82a2-c0f1ac85ac83","Type":"ContainerStarted","Data":"e2fa65261d812eab922d7916a26c3fb91cc85d1cd0457d905991377e1e2ba15f"} Dec 04 04:48:42 crc kubenswrapper[4806]: I1204 04:48:42.037161 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8ldlh" podStartSLOduration=3.6232836710000003 podStartE2EDuration="7.037145491s" podCreationTimestamp="2025-12-04 04:48:35 +0000 UTC" firstStartedPulling="2025-12-04 04:48:37.943142329 +0000 UTC m=+3232.801655257" lastFinishedPulling="2025-12-04 04:48:41.357004149 +0000 UTC m=+3236.215517077" observedRunningTime="2025-12-04 04:48:42.031488233 +0000 UTC m=+3236.890001161" watchObservedRunningTime="2025-12-04 04:48:42.037145491 +0000 UTC m=+3236.895658409" Dec 04 04:48:46 crc kubenswrapper[4806]: I1204 04:48:46.180966 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:46 crc kubenswrapper[4806]: I1204 04:48:46.181613 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:46 crc kubenswrapper[4806]: I1204 04:48:46.231820 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:47 crc kubenswrapper[4806]: I1204 04:48:47.099620 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:49 crc kubenswrapper[4806]: I1204 04:48:49.630650 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8ldlh"] Dec 04 04:48:49 crc kubenswrapper[4806]: I1204 04:48:49.634865 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8ldlh" podUID="2a187bb3-1801-491d-82a2-c0f1ac85ac83" containerName="registry-server" containerID="cri-o://e2fa65261d812eab922d7916a26c3fb91cc85d1cd0457d905991377e1e2ba15f" gracePeriod=2 Dec 04 04:48:50 crc kubenswrapper[4806]: I1204 04:48:50.070583 4806 generic.go:334] "Generic (PLEG): container finished" podID="2a187bb3-1801-491d-82a2-c0f1ac85ac83" containerID="e2fa65261d812eab922d7916a26c3fb91cc85d1cd0457d905991377e1e2ba15f" exitCode=0 Dec 04 04:48:50 crc kubenswrapper[4806]: I1204 04:48:50.070627 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ldlh" event={"ID":"2a187bb3-1801-491d-82a2-c0f1ac85ac83","Type":"ContainerDied","Data":"e2fa65261d812eab922d7916a26c3fb91cc85d1cd0457d905991377e1e2ba15f"} Dec 04 04:48:50 crc kubenswrapper[4806]: I1204 04:48:50.435587 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:50 crc kubenswrapper[4806]: I1204 04:48:50.506605 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a187bb3-1801-491d-82a2-c0f1ac85ac83-utilities\") pod \"2a187bb3-1801-491d-82a2-c0f1ac85ac83\" (UID: \"2a187bb3-1801-491d-82a2-c0f1ac85ac83\") " Dec 04 04:48:50 crc kubenswrapper[4806]: I1204 04:48:50.506686 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lw6fx\" (UniqueName: \"kubernetes.io/projected/2a187bb3-1801-491d-82a2-c0f1ac85ac83-kube-api-access-lw6fx\") pod \"2a187bb3-1801-491d-82a2-c0f1ac85ac83\" (UID: \"2a187bb3-1801-491d-82a2-c0f1ac85ac83\") " Dec 04 04:48:50 crc kubenswrapper[4806]: I1204 04:48:50.506862 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a187bb3-1801-491d-82a2-c0f1ac85ac83-catalog-content\") pod \"2a187bb3-1801-491d-82a2-c0f1ac85ac83\" (UID: \"2a187bb3-1801-491d-82a2-c0f1ac85ac83\") " Dec 04 04:48:50 crc kubenswrapper[4806]: I1204 04:48:50.507919 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a187bb3-1801-491d-82a2-c0f1ac85ac83-utilities" (OuterVolumeSpecName: "utilities") pod "2a187bb3-1801-491d-82a2-c0f1ac85ac83" (UID: "2a187bb3-1801-491d-82a2-c0f1ac85ac83"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:48:50 crc kubenswrapper[4806]: I1204 04:48:50.548651 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a187bb3-1801-491d-82a2-c0f1ac85ac83-kube-api-access-lw6fx" (OuterVolumeSpecName: "kube-api-access-lw6fx") pod "2a187bb3-1801-491d-82a2-c0f1ac85ac83" (UID: "2a187bb3-1801-491d-82a2-c0f1ac85ac83"). InnerVolumeSpecName "kube-api-access-lw6fx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:48:50 crc kubenswrapper[4806]: I1204 04:48:50.578265 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a187bb3-1801-491d-82a2-c0f1ac85ac83-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a187bb3-1801-491d-82a2-c0f1ac85ac83" (UID: "2a187bb3-1801-491d-82a2-c0f1ac85ac83"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:48:50 crc kubenswrapper[4806]: I1204 04:48:50.610008 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a187bb3-1801-491d-82a2-c0f1ac85ac83-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:48:50 crc kubenswrapper[4806]: I1204 04:48:50.610044 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lw6fx\" (UniqueName: \"kubernetes.io/projected/2a187bb3-1801-491d-82a2-c0f1ac85ac83-kube-api-access-lw6fx\") on node \"crc\" DevicePath \"\"" Dec 04 04:48:50 crc kubenswrapper[4806]: I1204 04:48:50.610054 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a187bb3-1801-491d-82a2-c0f1ac85ac83-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:48:51 crc kubenswrapper[4806]: I1204 04:48:51.080349 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8ldlh" event={"ID":"2a187bb3-1801-491d-82a2-c0f1ac85ac83","Type":"ContainerDied","Data":"44ab2e04a73418783dd34b4509e767a275a26f371a45bbc3986e1d681ab4be43"} Dec 04 04:48:51 crc kubenswrapper[4806]: I1204 04:48:51.080615 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8ldlh" Dec 04 04:48:51 crc kubenswrapper[4806]: I1204 04:48:51.080990 4806 scope.go:117] "RemoveContainer" containerID="e2fa65261d812eab922d7916a26c3fb91cc85d1cd0457d905991377e1e2ba15f" Dec 04 04:48:51 crc kubenswrapper[4806]: I1204 04:48:51.107095 4806 scope.go:117] "RemoveContainer" containerID="1e1c3b6fff95616d6c78a8dd542cce8a20af42c80f86d809a70a25023c999c85" Dec 04 04:48:51 crc kubenswrapper[4806]: I1204 04:48:51.121776 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8ldlh"] Dec 04 04:48:51 crc kubenswrapper[4806]: I1204 04:48:51.131460 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8ldlh"] Dec 04 04:48:51 crc kubenswrapper[4806]: I1204 04:48:51.151044 4806 scope.go:117] "RemoveContainer" containerID="be080212336c6ad4e59b26baff5d0fe23ac2bb77e541ace6a947e9e39916bf3d" Dec 04 04:48:51 crc kubenswrapper[4806]: I1204 04:48:51.435021 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a187bb3-1801-491d-82a2-c0f1ac85ac83" path="/var/lib/kubelet/pods/2a187bb3-1801-491d-82a2-c0f1ac85ac83/volumes" Dec 04 04:49:57 crc kubenswrapper[4806]: I1204 04:49:57.047609 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:49:57 crc kubenswrapper[4806]: I1204 04:49:57.048592 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.647565 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5qmxb"] Dec 04 04:50:00 crc kubenswrapper[4806]: E1204 04:50:00.648674 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a187bb3-1801-491d-82a2-c0f1ac85ac83" containerName="registry-server" Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.648694 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a187bb3-1801-491d-82a2-c0f1ac85ac83" containerName="registry-server" Dec 04 04:50:00 crc kubenswrapper[4806]: E1204 04:50:00.648717 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a187bb3-1801-491d-82a2-c0f1ac85ac83" containerName="extract-utilities" Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.648728 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a187bb3-1801-491d-82a2-c0f1ac85ac83" containerName="extract-utilities" Dec 04 04:50:00 crc kubenswrapper[4806]: E1204 04:50:00.648759 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a187bb3-1801-491d-82a2-c0f1ac85ac83" containerName="extract-content" Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.648769 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a187bb3-1801-491d-82a2-c0f1ac85ac83" containerName="extract-content" Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.650065 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a187bb3-1801-491d-82a2-c0f1ac85ac83" containerName="registry-server" Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.651942 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.674080 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5qmxb"] Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.742023 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ce53346-d426-4254-8c08-0065c6d78ce8-utilities\") pod \"redhat-operators-5qmxb\" (UID: \"0ce53346-d426-4254-8c08-0065c6d78ce8\") " pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.742085 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xswwt\" (UniqueName: \"kubernetes.io/projected/0ce53346-d426-4254-8c08-0065c6d78ce8-kube-api-access-xswwt\") pod \"redhat-operators-5qmxb\" (UID: \"0ce53346-d426-4254-8c08-0065c6d78ce8\") " pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.742209 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ce53346-d426-4254-8c08-0065c6d78ce8-catalog-content\") pod \"redhat-operators-5qmxb\" (UID: \"0ce53346-d426-4254-8c08-0065c6d78ce8\") " pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.843708 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ce53346-d426-4254-8c08-0065c6d78ce8-catalog-content\") pod \"redhat-operators-5qmxb\" (UID: \"0ce53346-d426-4254-8c08-0065c6d78ce8\") " pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.843868 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ce53346-d426-4254-8c08-0065c6d78ce8-utilities\") pod \"redhat-operators-5qmxb\" (UID: \"0ce53346-d426-4254-8c08-0065c6d78ce8\") " pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.843885 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xswwt\" (UniqueName: \"kubernetes.io/projected/0ce53346-d426-4254-8c08-0065c6d78ce8-kube-api-access-xswwt\") pod \"redhat-operators-5qmxb\" (UID: \"0ce53346-d426-4254-8c08-0065c6d78ce8\") " pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.844453 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ce53346-d426-4254-8c08-0065c6d78ce8-catalog-content\") pod \"redhat-operators-5qmxb\" (UID: \"0ce53346-d426-4254-8c08-0065c6d78ce8\") " pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.844471 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ce53346-d426-4254-8c08-0065c6d78ce8-utilities\") pod \"redhat-operators-5qmxb\" (UID: \"0ce53346-d426-4254-8c08-0065c6d78ce8\") " pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.881993 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xswwt\" (UniqueName: \"kubernetes.io/projected/0ce53346-d426-4254-8c08-0065c6d78ce8-kube-api-access-xswwt\") pod \"redhat-operators-5qmxb\" (UID: \"0ce53346-d426-4254-8c08-0065c6d78ce8\") " pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:00 crc kubenswrapper[4806]: I1204 04:50:00.984172 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:01 crc kubenswrapper[4806]: I1204 04:50:01.727868 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5qmxb"] Dec 04 04:50:02 crc kubenswrapper[4806]: I1204 04:50:02.711532 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5qmxb" event={"ID":"0ce53346-d426-4254-8c08-0065c6d78ce8","Type":"ContainerStarted","Data":"9074a27dbb91c5d8c92e57d614d708e41ab53c99f1f8e2a531caf84a3e3019de"} Dec 04 04:50:02 crc kubenswrapper[4806]: I1204 04:50:02.712129 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5qmxb" event={"ID":"0ce53346-d426-4254-8c08-0065c6d78ce8","Type":"ContainerStarted","Data":"13ac07bd63c8fb709d0747547c429c99a1aa7069018f3c2ce2ec4a1262acf115"} Dec 04 04:50:03 crc kubenswrapper[4806]: I1204 04:50:03.721038 4806 generic.go:334] "Generic (PLEG): container finished" podID="0ce53346-d426-4254-8c08-0065c6d78ce8" containerID="9074a27dbb91c5d8c92e57d614d708e41ab53c99f1f8e2a531caf84a3e3019de" exitCode=0 Dec 04 04:50:03 crc kubenswrapper[4806]: I1204 04:50:03.721089 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5qmxb" event={"ID":"0ce53346-d426-4254-8c08-0065c6d78ce8","Type":"ContainerDied","Data":"9074a27dbb91c5d8c92e57d614d708e41ab53c99f1f8e2a531caf84a3e3019de"} Dec 04 04:50:04 crc kubenswrapper[4806]: I1204 04:50:04.753651 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5qmxb" event={"ID":"0ce53346-d426-4254-8c08-0065c6d78ce8","Type":"ContainerStarted","Data":"fcd3e487640c52eda95545637e567b67395fea395e22f3290991d699dc7bbec1"} Dec 04 04:50:08 crc kubenswrapper[4806]: I1204 04:50:08.787479 4806 generic.go:334] "Generic (PLEG): container finished" podID="0ce53346-d426-4254-8c08-0065c6d78ce8" containerID="fcd3e487640c52eda95545637e567b67395fea395e22f3290991d699dc7bbec1" exitCode=0 Dec 04 04:50:08 crc kubenswrapper[4806]: I1204 04:50:08.787564 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5qmxb" event={"ID":"0ce53346-d426-4254-8c08-0065c6d78ce8","Type":"ContainerDied","Data":"fcd3e487640c52eda95545637e567b67395fea395e22f3290991d699dc7bbec1"} Dec 04 04:50:09 crc kubenswrapper[4806]: I1204 04:50:09.802264 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5qmxb" event={"ID":"0ce53346-d426-4254-8c08-0065c6d78ce8","Type":"ContainerStarted","Data":"3443fad88a1fd3e73d56d14b000471a8a15e3cbf6a6037d3cd2f995fa57ca3d1"} Dec 04 04:50:09 crc kubenswrapper[4806]: I1204 04:50:09.829588 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5qmxb" podStartSLOduration=4.297756543 podStartE2EDuration="9.829566558s" podCreationTimestamp="2025-12-04 04:50:00 +0000 UTC" firstStartedPulling="2025-12-04 04:50:03.723386539 +0000 UTC m=+3318.581899467" lastFinishedPulling="2025-12-04 04:50:09.255196554 +0000 UTC m=+3324.113709482" observedRunningTime="2025-12-04 04:50:09.821471333 +0000 UTC m=+3324.679984261" watchObservedRunningTime="2025-12-04 04:50:09.829566558 +0000 UTC m=+3324.688079486" Dec 04 04:50:10 crc kubenswrapper[4806]: I1204 04:50:10.984753 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:10 crc kubenswrapper[4806]: I1204 04:50:10.985097 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:12 crc kubenswrapper[4806]: I1204 04:50:12.033338 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5qmxb" podUID="0ce53346-d426-4254-8c08-0065c6d78ce8" containerName="registry-server" probeResult="failure" output=< Dec 04 04:50:12 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Dec 04 04:50:12 crc kubenswrapper[4806]: > Dec 04 04:50:22 crc kubenswrapper[4806]: I1204 04:50:22.039364 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5qmxb" podUID="0ce53346-d426-4254-8c08-0065c6d78ce8" containerName="registry-server" probeResult="failure" output=< Dec 04 04:50:22 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Dec 04 04:50:22 crc kubenswrapper[4806]: > Dec 04 04:50:27 crc kubenswrapper[4806]: I1204 04:50:27.047307 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:50:27 crc kubenswrapper[4806]: I1204 04:50:27.047905 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:50:31 crc kubenswrapper[4806]: I1204 04:50:31.036484 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:31 crc kubenswrapper[4806]: I1204 04:50:31.089263 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:31 crc kubenswrapper[4806]: I1204 04:50:31.874133 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5qmxb"] Dec 04 04:50:33 crc kubenswrapper[4806]: I1204 04:50:33.028087 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5qmxb" podUID="0ce53346-d426-4254-8c08-0065c6d78ce8" containerName="registry-server" containerID="cri-o://3443fad88a1fd3e73d56d14b000471a8a15e3cbf6a6037d3cd2f995fa57ca3d1" gracePeriod=2 Dec 04 04:50:33 crc kubenswrapper[4806]: I1204 04:50:33.827012 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:33 crc kubenswrapper[4806]: I1204 04:50:33.890971 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xswwt\" (UniqueName: \"kubernetes.io/projected/0ce53346-d426-4254-8c08-0065c6d78ce8-kube-api-access-xswwt\") pod \"0ce53346-d426-4254-8c08-0065c6d78ce8\" (UID: \"0ce53346-d426-4254-8c08-0065c6d78ce8\") " Dec 04 04:50:33 crc kubenswrapper[4806]: I1204 04:50:33.891087 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ce53346-d426-4254-8c08-0065c6d78ce8-catalog-content\") pod \"0ce53346-d426-4254-8c08-0065c6d78ce8\" (UID: \"0ce53346-d426-4254-8c08-0065c6d78ce8\") " Dec 04 04:50:33 crc kubenswrapper[4806]: I1204 04:50:33.891140 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ce53346-d426-4254-8c08-0065c6d78ce8-utilities\") pod \"0ce53346-d426-4254-8c08-0065c6d78ce8\" (UID: \"0ce53346-d426-4254-8c08-0065c6d78ce8\") " Dec 04 04:50:33 crc kubenswrapper[4806]: I1204 04:50:33.892070 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ce53346-d426-4254-8c08-0065c6d78ce8-utilities" (OuterVolumeSpecName: "utilities") pod "0ce53346-d426-4254-8c08-0065c6d78ce8" (UID: "0ce53346-d426-4254-8c08-0065c6d78ce8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:50:33 crc kubenswrapper[4806]: I1204 04:50:33.907348 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ce53346-d426-4254-8c08-0065c6d78ce8-kube-api-access-xswwt" (OuterVolumeSpecName: "kube-api-access-xswwt") pod "0ce53346-d426-4254-8c08-0065c6d78ce8" (UID: "0ce53346-d426-4254-8c08-0065c6d78ce8"). InnerVolumeSpecName "kube-api-access-xswwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:50:33 crc kubenswrapper[4806]: I1204 04:50:33.996176 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xswwt\" (UniqueName: \"kubernetes.io/projected/0ce53346-d426-4254-8c08-0065c6d78ce8-kube-api-access-xswwt\") on node \"crc\" DevicePath \"\"" Dec 04 04:50:33 crc kubenswrapper[4806]: I1204 04:50:33.996205 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ce53346-d426-4254-8c08-0065c6d78ce8-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.020009 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ce53346-d426-4254-8c08-0065c6d78ce8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ce53346-d426-4254-8c08-0065c6d78ce8" (UID: "0ce53346-d426-4254-8c08-0065c6d78ce8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.040811 4806 generic.go:334] "Generic (PLEG): container finished" podID="0ce53346-d426-4254-8c08-0065c6d78ce8" containerID="3443fad88a1fd3e73d56d14b000471a8a15e3cbf6a6037d3cd2f995fa57ca3d1" exitCode=0 Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.040875 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5qmxb" Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.040902 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5qmxb" event={"ID":"0ce53346-d426-4254-8c08-0065c6d78ce8","Type":"ContainerDied","Data":"3443fad88a1fd3e73d56d14b000471a8a15e3cbf6a6037d3cd2f995fa57ca3d1"} Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.041995 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5qmxb" event={"ID":"0ce53346-d426-4254-8c08-0065c6d78ce8","Type":"ContainerDied","Data":"13ac07bd63c8fb709d0747547c429c99a1aa7069018f3c2ce2ec4a1262acf115"} Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.042022 4806 scope.go:117] "RemoveContainer" containerID="3443fad88a1fd3e73d56d14b000471a8a15e3cbf6a6037d3cd2f995fa57ca3d1" Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.079319 4806 scope.go:117] "RemoveContainer" containerID="fcd3e487640c52eda95545637e567b67395fea395e22f3290991d699dc7bbec1" Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.121027 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ce53346-d426-4254-8c08-0065c6d78ce8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.122019 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5qmxb"] Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.134761 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5qmxb"] Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.146740 4806 scope.go:117] "RemoveContainer" containerID="9074a27dbb91c5d8c92e57d614d708e41ab53c99f1f8e2a531caf84a3e3019de" Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.207064 4806 scope.go:117] "RemoveContainer" containerID="3443fad88a1fd3e73d56d14b000471a8a15e3cbf6a6037d3cd2f995fa57ca3d1" Dec 04 04:50:34 crc kubenswrapper[4806]: E1204 04:50:34.208552 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3443fad88a1fd3e73d56d14b000471a8a15e3cbf6a6037d3cd2f995fa57ca3d1\": container with ID starting with 3443fad88a1fd3e73d56d14b000471a8a15e3cbf6a6037d3cd2f995fa57ca3d1 not found: ID does not exist" containerID="3443fad88a1fd3e73d56d14b000471a8a15e3cbf6a6037d3cd2f995fa57ca3d1" Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.208608 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3443fad88a1fd3e73d56d14b000471a8a15e3cbf6a6037d3cd2f995fa57ca3d1"} err="failed to get container status \"3443fad88a1fd3e73d56d14b000471a8a15e3cbf6a6037d3cd2f995fa57ca3d1\": rpc error: code = NotFound desc = could not find container \"3443fad88a1fd3e73d56d14b000471a8a15e3cbf6a6037d3cd2f995fa57ca3d1\": container with ID starting with 3443fad88a1fd3e73d56d14b000471a8a15e3cbf6a6037d3cd2f995fa57ca3d1 not found: ID does not exist" Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.208643 4806 scope.go:117] "RemoveContainer" containerID="fcd3e487640c52eda95545637e567b67395fea395e22f3290991d699dc7bbec1" Dec 04 04:50:34 crc kubenswrapper[4806]: E1204 04:50:34.212198 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcd3e487640c52eda95545637e567b67395fea395e22f3290991d699dc7bbec1\": container with ID starting with fcd3e487640c52eda95545637e567b67395fea395e22f3290991d699dc7bbec1 not found: ID does not exist" containerID="fcd3e487640c52eda95545637e567b67395fea395e22f3290991d699dc7bbec1" Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.212250 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcd3e487640c52eda95545637e567b67395fea395e22f3290991d699dc7bbec1"} err="failed to get container status \"fcd3e487640c52eda95545637e567b67395fea395e22f3290991d699dc7bbec1\": rpc error: code = NotFound desc = could not find container \"fcd3e487640c52eda95545637e567b67395fea395e22f3290991d699dc7bbec1\": container with ID starting with fcd3e487640c52eda95545637e567b67395fea395e22f3290991d699dc7bbec1 not found: ID does not exist" Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.212277 4806 scope.go:117] "RemoveContainer" containerID="9074a27dbb91c5d8c92e57d614d708e41ab53c99f1f8e2a531caf84a3e3019de" Dec 04 04:50:34 crc kubenswrapper[4806]: E1204 04:50:34.214727 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9074a27dbb91c5d8c92e57d614d708e41ab53c99f1f8e2a531caf84a3e3019de\": container with ID starting with 9074a27dbb91c5d8c92e57d614d708e41ab53c99f1f8e2a531caf84a3e3019de not found: ID does not exist" containerID="9074a27dbb91c5d8c92e57d614d708e41ab53c99f1f8e2a531caf84a3e3019de" Dec 04 04:50:34 crc kubenswrapper[4806]: I1204 04:50:34.214768 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9074a27dbb91c5d8c92e57d614d708e41ab53c99f1f8e2a531caf84a3e3019de"} err="failed to get container status \"9074a27dbb91c5d8c92e57d614d708e41ab53c99f1f8e2a531caf84a3e3019de\": rpc error: code = NotFound desc = could not find container \"9074a27dbb91c5d8c92e57d614d708e41ab53c99f1f8e2a531caf84a3e3019de\": container with ID starting with 9074a27dbb91c5d8c92e57d614d708e41ab53c99f1f8e2a531caf84a3e3019de not found: ID does not exist" Dec 04 04:50:35 crc kubenswrapper[4806]: I1204 04:50:35.434532 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ce53346-d426-4254-8c08-0065c6d78ce8" path="/var/lib/kubelet/pods/0ce53346-d426-4254-8c08-0065c6d78ce8/volumes" Dec 04 04:50:57 crc kubenswrapper[4806]: I1204 04:50:57.047393 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:50:57 crc kubenswrapper[4806]: I1204 04:50:57.048000 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:50:57 crc kubenswrapper[4806]: I1204 04:50:57.048047 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 04:50:57 crc kubenswrapper[4806]: I1204 04:50:57.048787 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"180699dbf0c533c23b37a9f63f83cd78ab2f3830f234d8427f5b5bc67b2a3005"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 04:50:57 crc kubenswrapper[4806]: I1204 04:50:57.048841 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://180699dbf0c533c23b37a9f63f83cd78ab2f3830f234d8427f5b5bc67b2a3005" gracePeriod=600 Dec 04 04:50:57 crc kubenswrapper[4806]: I1204 04:50:57.298555 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="180699dbf0c533c23b37a9f63f83cd78ab2f3830f234d8427f5b5bc67b2a3005" exitCode=0 Dec 04 04:50:57 crc kubenswrapper[4806]: I1204 04:50:57.298955 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"180699dbf0c533c23b37a9f63f83cd78ab2f3830f234d8427f5b5bc67b2a3005"} Dec 04 04:50:57 crc kubenswrapper[4806]: I1204 04:50:57.298997 4806 scope.go:117] "RemoveContainer" containerID="1254c4fbf1ccb2ca57c552e7b25a98a20f4347e316f961925bab1ecdad2f282f" Dec 04 04:50:58 crc kubenswrapper[4806]: I1204 04:50:58.311872 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47"} Dec 04 04:52:57 crc kubenswrapper[4806]: I1204 04:52:57.046787 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:52:57 crc kubenswrapper[4806]: I1204 04:52:57.047341 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:53:27 crc kubenswrapper[4806]: I1204 04:53:27.047176 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:53:27 crc kubenswrapper[4806]: I1204 04:53:27.048856 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.090056 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wdtbt"] Dec 04 04:53:48 crc kubenswrapper[4806]: E1204 04:53:48.092185 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ce53346-d426-4254-8c08-0065c6d78ce8" containerName="extract-content" Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.092221 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ce53346-d426-4254-8c08-0065c6d78ce8" containerName="extract-content" Dec 04 04:53:48 crc kubenswrapper[4806]: E1204 04:53:48.092242 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ce53346-d426-4254-8c08-0065c6d78ce8" containerName="extract-utilities" Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.092250 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ce53346-d426-4254-8c08-0065c6d78ce8" containerName="extract-utilities" Dec 04 04:53:48 crc kubenswrapper[4806]: E1204 04:53:48.092286 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ce53346-d426-4254-8c08-0065c6d78ce8" containerName="registry-server" Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.092299 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ce53346-d426-4254-8c08-0065c6d78ce8" containerName="registry-server" Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.092515 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ce53346-d426-4254-8c08-0065c6d78ce8" containerName="registry-server" Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.094251 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.181703 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wdtbt"] Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.259759 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb77b628-af1b-4015-a305-32292d64a553-catalog-content\") pod \"community-operators-wdtbt\" (UID: \"bb77b628-af1b-4015-a305-32292d64a553\") " pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.259828 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb77b628-af1b-4015-a305-32292d64a553-utilities\") pod \"community-operators-wdtbt\" (UID: \"bb77b628-af1b-4015-a305-32292d64a553\") " pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.259879 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pswg9\" (UniqueName: \"kubernetes.io/projected/bb77b628-af1b-4015-a305-32292d64a553-kube-api-access-pswg9\") pod \"community-operators-wdtbt\" (UID: \"bb77b628-af1b-4015-a305-32292d64a553\") " pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.361780 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb77b628-af1b-4015-a305-32292d64a553-catalog-content\") pod \"community-operators-wdtbt\" (UID: \"bb77b628-af1b-4015-a305-32292d64a553\") " pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.361856 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb77b628-af1b-4015-a305-32292d64a553-utilities\") pod \"community-operators-wdtbt\" (UID: \"bb77b628-af1b-4015-a305-32292d64a553\") " pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.361891 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pswg9\" (UniqueName: \"kubernetes.io/projected/bb77b628-af1b-4015-a305-32292d64a553-kube-api-access-pswg9\") pod \"community-operators-wdtbt\" (UID: \"bb77b628-af1b-4015-a305-32292d64a553\") " pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.362476 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb77b628-af1b-4015-a305-32292d64a553-utilities\") pod \"community-operators-wdtbt\" (UID: \"bb77b628-af1b-4015-a305-32292d64a553\") " pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.362538 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb77b628-af1b-4015-a305-32292d64a553-catalog-content\") pod \"community-operators-wdtbt\" (UID: \"bb77b628-af1b-4015-a305-32292d64a553\") " pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.386348 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pswg9\" (UniqueName: \"kubernetes.io/projected/bb77b628-af1b-4015-a305-32292d64a553-kube-api-access-pswg9\") pod \"community-operators-wdtbt\" (UID: \"bb77b628-af1b-4015-a305-32292d64a553\") " pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:53:48 crc kubenswrapper[4806]: I1204 04:53:48.412153 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:53:49 crc kubenswrapper[4806]: I1204 04:53:49.112478 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wdtbt"] Dec 04 04:53:50 crc kubenswrapper[4806]: I1204 04:53:49.999779 4806 generic.go:334] "Generic (PLEG): container finished" podID="bb77b628-af1b-4015-a305-32292d64a553" containerID="31c8c52f613d030396e3232137da6cde4871ca350b3112369b28197792bbe2e4" exitCode=0 Dec 04 04:53:50 crc kubenswrapper[4806]: I1204 04:53:49.999980 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdtbt" event={"ID":"bb77b628-af1b-4015-a305-32292d64a553","Type":"ContainerDied","Data":"31c8c52f613d030396e3232137da6cde4871ca350b3112369b28197792bbe2e4"} Dec 04 04:53:50 crc kubenswrapper[4806]: I1204 04:53:50.000105 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdtbt" event={"ID":"bb77b628-af1b-4015-a305-32292d64a553","Type":"ContainerStarted","Data":"5fa593476e5733539637134b5d6673f97e044748515ccba5f63150e82aee5419"} Dec 04 04:53:50 crc kubenswrapper[4806]: I1204 04:53:50.003178 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 04:53:51 crc kubenswrapper[4806]: I1204 04:53:51.013350 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdtbt" event={"ID":"bb77b628-af1b-4015-a305-32292d64a553","Type":"ContainerStarted","Data":"912294280f99d80041d2c035c6c28f9a3b9aae67d5742d8820de3351db533217"} Dec 04 04:53:52 crc kubenswrapper[4806]: I1204 04:53:52.023486 4806 generic.go:334] "Generic (PLEG): container finished" podID="bb77b628-af1b-4015-a305-32292d64a553" containerID="912294280f99d80041d2c035c6c28f9a3b9aae67d5742d8820de3351db533217" exitCode=0 Dec 04 04:53:52 crc kubenswrapper[4806]: I1204 04:53:52.023528 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdtbt" event={"ID":"bb77b628-af1b-4015-a305-32292d64a553","Type":"ContainerDied","Data":"912294280f99d80041d2c035c6c28f9a3b9aae67d5742d8820de3351db533217"} Dec 04 04:53:53 crc kubenswrapper[4806]: I1204 04:53:53.034950 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdtbt" event={"ID":"bb77b628-af1b-4015-a305-32292d64a553","Type":"ContainerStarted","Data":"cb9c69372b7785e69feebf6ed4a8e538bf71951d82cb75fb4978294079d4bcbc"} Dec 04 04:53:53 crc kubenswrapper[4806]: I1204 04:53:53.061948 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wdtbt" podStartSLOduration=2.585993581 podStartE2EDuration="5.061907337s" podCreationTimestamp="2025-12-04 04:53:48 +0000 UTC" firstStartedPulling="2025-12-04 04:53:50.00280269 +0000 UTC m=+3544.861315618" lastFinishedPulling="2025-12-04 04:53:52.478716446 +0000 UTC m=+3547.337229374" observedRunningTime="2025-12-04 04:53:53.055759714 +0000 UTC m=+3547.914272642" watchObservedRunningTime="2025-12-04 04:53:53.061907337 +0000 UTC m=+3547.920420265" Dec 04 04:53:57 crc kubenswrapper[4806]: I1204 04:53:57.047226 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 04:53:57 crc kubenswrapper[4806]: I1204 04:53:57.047581 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 04:53:57 crc kubenswrapper[4806]: I1204 04:53:57.047622 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 04:53:57 crc kubenswrapper[4806]: I1204 04:53:57.048382 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 04:53:57 crc kubenswrapper[4806]: I1204 04:53:57.048445 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" gracePeriod=600 Dec 04 04:53:57 crc kubenswrapper[4806]: E1204 04:53:57.168794 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:53:58 crc kubenswrapper[4806]: I1204 04:53:58.095299 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" exitCode=0 Dec 04 04:53:58 crc kubenswrapper[4806]: I1204 04:53:58.095371 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47"} Dec 04 04:53:58 crc kubenswrapper[4806]: I1204 04:53:58.095660 4806 scope.go:117] "RemoveContainer" containerID="180699dbf0c533c23b37a9f63f83cd78ab2f3830f234d8427f5b5bc67b2a3005" Dec 04 04:53:58 crc kubenswrapper[4806]: I1204 04:53:58.096378 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:53:58 crc kubenswrapper[4806]: E1204 04:53:58.096775 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:53:58 crc kubenswrapper[4806]: I1204 04:53:58.412979 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:53:58 crc kubenswrapper[4806]: I1204 04:53:58.413077 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:53:58 crc kubenswrapper[4806]: I1204 04:53:58.470417 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:53:59 crc kubenswrapper[4806]: I1204 04:53:59.203830 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:53:59 crc kubenswrapper[4806]: I1204 04:53:59.311163 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wdtbt"] Dec 04 04:54:01 crc kubenswrapper[4806]: I1204 04:54:01.127496 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wdtbt" podUID="bb77b628-af1b-4015-a305-32292d64a553" containerName="registry-server" containerID="cri-o://cb9c69372b7785e69feebf6ed4a8e538bf71951d82cb75fb4978294079d4bcbc" gracePeriod=2 Dec 04 04:54:01 crc kubenswrapper[4806]: E1204 04:54:01.519805 4806 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb77b628_af1b_4015_a305_32292d64a553.slice/crio-conmon-cb9c69372b7785e69feebf6ed4a8e538bf71951d82cb75fb4978294079d4bcbc.scope\": RecentStats: unable to find data in memory cache]" Dec 04 04:54:02 crc kubenswrapper[4806]: I1204 04:54:02.136953 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdtbt" event={"ID":"bb77b628-af1b-4015-a305-32292d64a553","Type":"ContainerDied","Data":"cb9c69372b7785e69feebf6ed4a8e538bf71951d82cb75fb4978294079d4bcbc"} Dec 04 04:54:02 crc kubenswrapper[4806]: I1204 04:54:02.136922 4806 generic.go:334] "Generic (PLEG): container finished" podID="bb77b628-af1b-4015-a305-32292d64a553" containerID="cb9c69372b7785e69feebf6ed4a8e538bf71951d82cb75fb4978294079d4bcbc" exitCode=0 Dec 04 04:54:02 crc kubenswrapper[4806]: I1204 04:54:02.137396 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wdtbt" event={"ID":"bb77b628-af1b-4015-a305-32292d64a553","Type":"ContainerDied","Data":"5fa593476e5733539637134b5d6673f97e044748515ccba5f63150e82aee5419"} Dec 04 04:54:02 crc kubenswrapper[4806]: I1204 04:54:02.137413 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fa593476e5733539637134b5d6673f97e044748515ccba5f63150e82aee5419" Dec 04 04:54:02 crc kubenswrapper[4806]: I1204 04:54:02.221100 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:54:02 crc kubenswrapper[4806]: I1204 04:54:02.368178 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb77b628-af1b-4015-a305-32292d64a553-utilities\") pod \"bb77b628-af1b-4015-a305-32292d64a553\" (UID: \"bb77b628-af1b-4015-a305-32292d64a553\") " Dec 04 04:54:02 crc kubenswrapper[4806]: I1204 04:54:02.368386 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb77b628-af1b-4015-a305-32292d64a553-catalog-content\") pod \"bb77b628-af1b-4015-a305-32292d64a553\" (UID: \"bb77b628-af1b-4015-a305-32292d64a553\") " Dec 04 04:54:02 crc kubenswrapper[4806]: I1204 04:54:02.368415 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pswg9\" (UniqueName: \"kubernetes.io/projected/bb77b628-af1b-4015-a305-32292d64a553-kube-api-access-pswg9\") pod \"bb77b628-af1b-4015-a305-32292d64a553\" (UID: \"bb77b628-af1b-4015-a305-32292d64a553\") " Dec 04 04:54:02 crc kubenswrapper[4806]: I1204 04:54:02.369274 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb77b628-af1b-4015-a305-32292d64a553-utilities" (OuterVolumeSpecName: "utilities") pod "bb77b628-af1b-4015-a305-32292d64a553" (UID: "bb77b628-af1b-4015-a305-32292d64a553"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:54:02 crc kubenswrapper[4806]: I1204 04:54:02.379239 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb77b628-af1b-4015-a305-32292d64a553-kube-api-access-pswg9" (OuterVolumeSpecName: "kube-api-access-pswg9") pod "bb77b628-af1b-4015-a305-32292d64a553" (UID: "bb77b628-af1b-4015-a305-32292d64a553"). InnerVolumeSpecName "kube-api-access-pswg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:54:02 crc kubenswrapper[4806]: I1204 04:54:02.424375 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb77b628-af1b-4015-a305-32292d64a553-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb77b628-af1b-4015-a305-32292d64a553" (UID: "bb77b628-af1b-4015-a305-32292d64a553"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:54:02 crc kubenswrapper[4806]: I1204 04:54:02.470586 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb77b628-af1b-4015-a305-32292d64a553-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:54:02 crc kubenswrapper[4806]: I1204 04:54:02.470815 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb77b628-af1b-4015-a305-32292d64a553-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:54:02 crc kubenswrapper[4806]: I1204 04:54:02.470911 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pswg9\" (UniqueName: \"kubernetes.io/projected/bb77b628-af1b-4015-a305-32292d64a553-kube-api-access-pswg9\") on node \"crc\" DevicePath \"\"" Dec 04 04:54:03 crc kubenswrapper[4806]: I1204 04:54:03.145218 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wdtbt" Dec 04 04:54:03 crc kubenswrapper[4806]: I1204 04:54:03.178952 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wdtbt"] Dec 04 04:54:03 crc kubenswrapper[4806]: I1204 04:54:03.190448 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wdtbt"] Dec 04 04:54:03 crc kubenswrapper[4806]: I1204 04:54:03.435670 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb77b628-af1b-4015-a305-32292d64a553" path="/var/lib/kubelet/pods/bb77b628-af1b-4015-a305-32292d64a553/volumes" Dec 04 04:54:08 crc kubenswrapper[4806]: I1204 04:54:08.424015 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:54:08 crc kubenswrapper[4806]: E1204 04:54:08.424742 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:54:20 crc kubenswrapper[4806]: I1204 04:54:20.424187 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:54:20 crc kubenswrapper[4806]: E1204 04:54:20.424910 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.165454 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-txlxl"] Dec 04 04:54:29 crc kubenswrapper[4806]: E1204 04:54:29.180217 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb77b628-af1b-4015-a305-32292d64a553" containerName="extract-utilities" Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.180272 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb77b628-af1b-4015-a305-32292d64a553" containerName="extract-utilities" Dec 04 04:54:29 crc kubenswrapper[4806]: E1204 04:54:29.180320 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb77b628-af1b-4015-a305-32292d64a553" containerName="registry-server" Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.180331 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb77b628-af1b-4015-a305-32292d64a553" containerName="registry-server" Dec 04 04:54:29 crc kubenswrapper[4806]: E1204 04:54:29.180372 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb77b628-af1b-4015-a305-32292d64a553" containerName="extract-content" Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.180388 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb77b628-af1b-4015-a305-32292d64a553" containerName="extract-content" Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.189537 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb77b628-af1b-4015-a305-32292d64a553" containerName="registry-server" Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.207686 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.240835 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-txlxl"] Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.279206 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfe20ff0-b434-4f25-98b2-b50b6de30fca-catalog-content\") pod \"redhat-marketplace-txlxl\" (UID: \"dfe20ff0-b434-4f25-98b2-b50b6de30fca\") " pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.279275 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfe20ff0-b434-4f25-98b2-b50b6de30fca-utilities\") pod \"redhat-marketplace-txlxl\" (UID: \"dfe20ff0-b434-4f25-98b2-b50b6de30fca\") " pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.279338 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqj5z\" (UniqueName: \"kubernetes.io/projected/dfe20ff0-b434-4f25-98b2-b50b6de30fca-kube-api-access-pqj5z\") pod \"redhat-marketplace-txlxl\" (UID: \"dfe20ff0-b434-4f25-98b2-b50b6de30fca\") " pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.381112 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfe20ff0-b434-4f25-98b2-b50b6de30fca-catalog-content\") pod \"redhat-marketplace-txlxl\" (UID: \"dfe20ff0-b434-4f25-98b2-b50b6de30fca\") " pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.381156 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfe20ff0-b434-4f25-98b2-b50b6de30fca-utilities\") pod \"redhat-marketplace-txlxl\" (UID: \"dfe20ff0-b434-4f25-98b2-b50b6de30fca\") " pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.381201 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqj5z\" (UniqueName: \"kubernetes.io/projected/dfe20ff0-b434-4f25-98b2-b50b6de30fca-kube-api-access-pqj5z\") pod \"redhat-marketplace-txlxl\" (UID: \"dfe20ff0-b434-4f25-98b2-b50b6de30fca\") " pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.381693 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfe20ff0-b434-4f25-98b2-b50b6de30fca-catalog-content\") pod \"redhat-marketplace-txlxl\" (UID: \"dfe20ff0-b434-4f25-98b2-b50b6de30fca\") " pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.381982 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfe20ff0-b434-4f25-98b2-b50b6de30fca-utilities\") pod \"redhat-marketplace-txlxl\" (UID: \"dfe20ff0-b434-4f25-98b2-b50b6de30fca\") " pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.412812 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqj5z\" (UniqueName: \"kubernetes.io/projected/dfe20ff0-b434-4f25-98b2-b50b6de30fca-kube-api-access-pqj5z\") pod \"redhat-marketplace-txlxl\" (UID: \"dfe20ff0-b434-4f25-98b2-b50b6de30fca\") " pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:29 crc kubenswrapper[4806]: I1204 04:54:29.554886 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:30 crc kubenswrapper[4806]: I1204 04:54:30.066490 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-txlxl"] Dec 04 04:54:30 crc kubenswrapper[4806]: I1204 04:54:30.390853 4806 generic.go:334] "Generic (PLEG): container finished" podID="dfe20ff0-b434-4f25-98b2-b50b6de30fca" containerID="26e90b0d9b58fc5422659edbff904cb5264dc49bbad1d141ed78b4042aa15298" exitCode=0 Dec 04 04:54:30 crc kubenswrapper[4806]: I1204 04:54:30.390903 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txlxl" event={"ID":"dfe20ff0-b434-4f25-98b2-b50b6de30fca","Type":"ContainerDied","Data":"26e90b0d9b58fc5422659edbff904cb5264dc49bbad1d141ed78b4042aa15298"} Dec 04 04:54:30 crc kubenswrapper[4806]: I1204 04:54:30.391064 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txlxl" event={"ID":"dfe20ff0-b434-4f25-98b2-b50b6de30fca","Type":"ContainerStarted","Data":"58db69e00b2f9ea0fa8f6e7d067f21227a3c178d841bdaca5b1813c4cb5b7bd6"} Dec 04 04:54:32 crc kubenswrapper[4806]: I1204 04:54:32.410201 4806 generic.go:334] "Generic (PLEG): container finished" podID="dfe20ff0-b434-4f25-98b2-b50b6de30fca" containerID="5388ce98e2c1a3cdc8afa4f9f0d12ffd7f57bca1694b8f473006c6eeb4644b02" exitCode=0 Dec 04 04:54:32 crc kubenswrapper[4806]: I1204 04:54:32.410787 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txlxl" event={"ID":"dfe20ff0-b434-4f25-98b2-b50b6de30fca","Type":"ContainerDied","Data":"5388ce98e2c1a3cdc8afa4f9f0d12ffd7f57bca1694b8f473006c6eeb4644b02"} Dec 04 04:54:33 crc kubenswrapper[4806]: I1204 04:54:33.423746 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:54:33 crc kubenswrapper[4806]: E1204 04:54:33.424837 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:54:33 crc kubenswrapper[4806]: I1204 04:54:33.435477 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txlxl" event={"ID":"dfe20ff0-b434-4f25-98b2-b50b6de30fca","Type":"ContainerStarted","Data":"78ea5df671a33ff566008234cdce1f7a6e769cafd883ef81d7b2e446992a271c"} Dec 04 04:54:39 crc kubenswrapper[4806]: I1204 04:54:39.555865 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:39 crc kubenswrapper[4806]: I1204 04:54:39.556520 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:39 crc kubenswrapper[4806]: I1204 04:54:39.607396 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:39 crc kubenswrapper[4806]: I1204 04:54:39.634317 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-txlxl" podStartSLOduration=8.220420191 podStartE2EDuration="10.634299796s" podCreationTimestamp="2025-12-04 04:54:29 +0000 UTC" firstStartedPulling="2025-12-04 04:54:30.392712888 +0000 UTC m=+3585.251225816" lastFinishedPulling="2025-12-04 04:54:32.806592493 +0000 UTC m=+3587.665105421" observedRunningTime="2025-12-04 04:54:33.457082292 +0000 UTC m=+3588.315595230" watchObservedRunningTime="2025-12-04 04:54:39.634299796 +0000 UTC m=+3594.492812884" Dec 04 04:54:40 crc kubenswrapper[4806]: I1204 04:54:40.537754 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:40 crc kubenswrapper[4806]: I1204 04:54:40.584224 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-txlxl"] Dec 04 04:54:42 crc kubenswrapper[4806]: I1204 04:54:42.500873 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-txlxl" podUID="dfe20ff0-b434-4f25-98b2-b50b6de30fca" containerName="registry-server" containerID="cri-o://78ea5df671a33ff566008234cdce1f7a6e769cafd883ef81d7b2e446992a271c" gracePeriod=2 Dec 04 04:54:42 crc kubenswrapper[4806]: E1204 04:54:42.598152 4806 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfe20ff0_b434_4f25_98b2_b50b6de30fca.slice/crio-conmon-78ea5df671a33ff566008234cdce1f7a6e769cafd883ef81d7b2e446992a271c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfe20ff0_b434_4f25_98b2_b50b6de30fca.slice/crio-78ea5df671a33ff566008234cdce1f7a6e769cafd883ef81d7b2e446992a271c.scope\": RecentStats: unable to find data in memory cache]" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.107530 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.187904 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqj5z\" (UniqueName: \"kubernetes.io/projected/dfe20ff0-b434-4f25-98b2-b50b6de30fca-kube-api-access-pqj5z\") pod \"dfe20ff0-b434-4f25-98b2-b50b6de30fca\" (UID: \"dfe20ff0-b434-4f25-98b2-b50b6de30fca\") " Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.188038 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfe20ff0-b434-4f25-98b2-b50b6de30fca-utilities\") pod \"dfe20ff0-b434-4f25-98b2-b50b6de30fca\" (UID: \"dfe20ff0-b434-4f25-98b2-b50b6de30fca\") " Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.188100 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfe20ff0-b434-4f25-98b2-b50b6de30fca-catalog-content\") pod \"dfe20ff0-b434-4f25-98b2-b50b6de30fca\" (UID: \"dfe20ff0-b434-4f25-98b2-b50b6de30fca\") " Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.189313 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfe20ff0-b434-4f25-98b2-b50b6de30fca-utilities" (OuterVolumeSpecName: "utilities") pod "dfe20ff0-b434-4f25-98b2-b50b6de30fca" (UID: "dfe20ff0-b434-4f25-98b2-b50b6de30fca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.194157 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfe20ff0-b434-4f25-98b2-b50b6de30fca-kube-api-access-pqj5z" (OuterVolumeSpecName: "kube-api-access-pqj5z") pod "dfe20ff0-b434-4f25-98b2-b50b6de30fca" (UID: "dfe20ff0-b434-4f25-98b2-b50b6de30fca"). InnerVolumeSpecName "kube-api-access-pqj5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.226219 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfe20ff0-b434-4f25-98b2-b50b6de30fca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dfe20ff0-b434-4f25-98b2-b50b6de30fca" (UID: "dfe20ff0-b434-4f25-98b2-b50b6de30fca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.290752 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqj5z\" (UniqueName: \"kubernetes.io/projected/dfe20ff0-b434-4f25-98b2-b50b6de30fca-kube-api-access-pqj5z\") on node \"crc\" DevicePath \"\"" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.290800 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfe20ff0-b434-4f25-98b2-b50b6de30fca-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.290817 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfe20ff0-b434-4f25-98b2-b50b6de30fca-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.513895 4806 generic.go:334] "Generic (PLEG): container finished" podID="dfe20ff0-b434-4f25-98b2-b50b6de30fca" containerID="78ea5df671a33ff566008234cdce1f7a6e769cafd883ef81d7b2e446992a271c" exitCode=0 Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.514030 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-txlxl" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.515517 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txlxl" event={"ID":"dfe20ff0-b434-4f25-98b2-b50b6de30fca","Type":"ContainerDied","Data":"78ea5df671a33ff566008234cdce1f7a6e769cafd883ef81d7b2e446992a271c"} Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.515620 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txlxl" event={"ID":"dfe20ff0-b434-4f25-98b2-b50b6de30fca","Type":"ContainerDied","Data":"58db69e00b2f9ea0fa8f6e7d067f21227a3c178d841bdaca5b1813c4cb5b7bd6"} Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.515693 4806 scope.go:117] "RemoveContainer" containerID="78ea5df671a33ff566008234cdce1f7a6e769cafd883ef81d7b2e446992a271c" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.543165 4806 scope.go:117] "RemoveContainer" containerID="5388ce98e2c1a3cdc8afa4f9f0d12ffd7f57bca1694b8f473006c6eeb4644b02" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.549996 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-txlxl"] Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.558037 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-txlxl"] Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.587161 4806 scope.go:117] "RemoveContainer" containerID="26e90b0d9b58fc5422659edbff904cb5264dc49bbad1d141ed78b4042aa15298" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.655331 4806 scope.go:117] "RemoveContainer" containerID="78ea5df671a33ff566008234cdce1f7a6e769cafd883ef81d7b2e446992a271c" Dec 04 04:54:43 crc kubenswrapper[4806]: E1204 04:54:43.657784 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78ea5df671a33ff566008234cdce1f7a6e769cafd883ef81d7b2e446992a271c\": container with ID starting with 78ea5df671a33ff566008234cdce1f7a6e769cafd883ef81d7b2e446992a271c not found: ID does not exist" containerID="78ea5df671a33ff566008234cdce1f7a6e769cafd883ef81d7b2e446992a271c" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.657819 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78ea5df671a33ff566008234cdce1f7a6e769cafd883ef81d7b2e446992a271c"} err="failed to get container status \"78ea5df671a33ff566008234cdce1f7a6e769cafd883ef81d7b2e446992a271c\": rpc error: code = NotFound desc = could not find container \"78ea5df671a33ff566008234cdce1f7a6e769cafd883ef81d7b2e446992a271c\": container with ID starting with 78ea5df671a33ff566008234cdce1f7a6e769cafd883ef81d7b2e446992a271c not found: ID does not exist" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.657843 4806 scope.go:117] "RemoveContainer" containerID="5388ce98e2c1a3cdc8afa4f9f0d12ffd7f57bca1694b8f473006c6eeb4644b02" Dec 04 04:54:43 crc kubenswrapper[4806]: E1204 04:54:43.658198 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5388ce98e2c1a3cdc8afa4f9f0d12ffd7f57bca1694b8f473006c6eeb4644b02\": container with ID starting with 5388ce98e2c1a3cdc8afa4f9f0d12ffd7f57bca1694b8f473006c6eeb4644b02 not found: ID does not exist" containerID="5388ce98e2c1a3cdc8afa4f9f0d12ffd7f57bca1694b8f473006c6eeb4644b02" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.658220 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5388ce98e2c1a3cdc8afa4f9f0d12ffd7f57bca1694b8f473006c6eeb4644b02"} err="failed to get container status \"5388ce98e2c1a3cdc8afa4f9f0d12ffd7f57bca1694b8f473006c6eeb4644b02\": rpc error: code = NotFound desc = could not find container \"5388ce98e2c1a3cdc8afa4f9f0d12ffd7f57bca1694b8f473006c6eeb4644b02\": container with ID starting with 5388ce98e2c1a3cdc8afa4f9f0d12ffd7f57bca1694b8f473006c6eeb4644b02 not found: ID does not exist" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.658239 4806 scope.go:117] "RemoveContainer" containerID="26e90b0d9b58fc5422659edbff904cb5264dc49bbad1d141ed78b4042aa15298" Dec 04 04:54:43 crc kubenswrapper[4806]: E1204 04:54:43.658810 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26e90b0d9b58fc5422659edbff904cb5264dc49bbad1d141ed78b4042aa15298\": container with ID starting with 26e90b0d9b58fc5422659edbff904cb5264dc49bbad1d141ed78b4042aa15298 not found: ID does not exist" containerID="26e90b0d9b58fc5422659edbff904cb5264dc49bbad1d141ed78b4042aa15298" Dec 04 04:54:43 crc kubenswrapper[4806]: I1204 04:54:43.658860 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26e90b0d9b58fc5422659edbff904cb5264dc49bbad1d141ed78b4042aa15298"} err="failed to get container status \"26e90b0d9b58fc5422659edbff904cb5264dc49bbad1d141ed78b4042aa15298\": rpc error: code = NotFound desc = could not find container \"26e90b0d9b58fc5422659edbff904cb5264dc49bbad1d141ed78b4042aa15298\": container with ID starting with 26e90b0d9b58fc5422659edbff904cb5264dc49bbad1d141ed78b4042aa15298 not found: ID does not exist" Dec 04 04:54:44 crc kubenswrapper[4806]: I1204 04:54:44.424411 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:54:44 crc kubenswrapper[4806]: E1204 04:54:44.424716 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:54:45 crc kubenswrapper[4806]: I1204 04:54:45.448145 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfe20ff0-b434-4f25-98b2-b50b6de30fca" path="/var/lib/kubelet/pods/dfe20ff0-b434-4f25-98b2-b50b6de30fca/volumes" Dec 04 04:54:59 crc kubenswrapper[4806]: I1204 04:54:59.424401 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:54:59 crc kubenswrapper[4806]: E1204 04:54:59.425265 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:55:11 crc kubenswrapper[4806]: I1204 04:55:11.425350 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:55:11 crc kubenswrapper[4806]: E1204 04:55:11.428280 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:55:26 crc kubenswrapper[4806]: I1204 04:55:26.424261 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:55:26 crc kubenswrapper[4806]: E1204 04:55:26.425055 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:55:37 crc kubenswrapper[4806]: I1204 04:55:37.424251 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:55:37 crc kubenswrapper[4806]: E1204 04:55:37.425081 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:55:52 crc kubenswrapper[4806]: I1204 04:55:52.423452 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:55:52 crc kubenswrapper[4806]: E1204 04:55:52.424341 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:56:06 crc kubenswrapper[4806]: I1204 04:56:06.423881 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:56:06 crc kubenswrapper[4806]: E1204 04:56:06.424704 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:56:17 crc kubenswrapper[4806]: I1204 04:56:17.423547 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:56:17 crc kubenswrapper[4806]: E1204 04:56:17.425711 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:56:28 crc kubenswrapper[4806]: I1204 04:56:28.423591 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:56:28 crc kubenswrapper[4806]: E1204 04:56:28.424348 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:56:40 crc kubenswrapper[4806]: I1204 04:56:40.424081 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:56:40 crc kubenswrapper[4806]: E1204 04:56:40.425410 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:56:52 crc kubenswrapper[4806]: I1204 04:56:52.424382 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:56:52 crc kubenswrapper[4806]: E1204 04:56:52.425081 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:57:06 crc kubenswrapper[4806]: I1204 04:57:06.424397 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:57:06 crc kubenswrapper[4806]: E1204 04:57:06.425172 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:57:17 crc kubenswrapper[4806]: I1204 04:57:17.424251 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:57:17 crc kubenswrapper[4806]: E1204 04:57:17.425204 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:57:29 crc kubenswrapper[4806]: I1204 04:57:29.424302 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:57:29 crc kubenswrapper[4806]: E1204 04:57:29.425051 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:57:40 crc kubenswrapper[4806]: I1204 04:57:40.423882 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:57:40 crc kubenswrapper[4806]: E1204 04:57:40.426249 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:57:53 crc kubenswrapper[4806]: I1204 04:57:53.423819 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:57:53 crc kubenswrapper[4806]: E1204 04:57:53.424778 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:58:05 crc kubenswrapper[4806]: I1204 04:58:05.433659 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:58:05 crc kubenswrapper[4806]: E1204 04:58:05.434449 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:58:20 crc kubenswrapper[4806]: I1204 04:58:20.423753 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:58:20 crc kubenswrapper[4806]: E1204 04:58:20.424503 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:58:35 crc kubenswrapper[4806]: I1204 04:58:35.432624 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:58:35 crc kubenswrapper[4806]: E1204 04:58:35.433414 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:58:49 crc kubenswrapper[4806]: I1204 04:58:49.423726 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:58:49 crc kubenswrapper[4806]: E1204 04:58:49.424551 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 04:59:04 crc kubenswrapper[4806]: I1204 04:59:04.423382 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 04:59:04 crc kubenswrapper[4806]: I1204 04:59:04.913360 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"b7a929db250bf8fd36191354c0529747a23a7c22591c6919b0c88de706949742"} Dec 04 04:59:25 crc kubenswrapper[4806]: I1204 04:59:25.942616 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-x8hzn"] Dec 04 04:59:25 crc kubenswrapper[4806]: E1204 04:59:25.943466 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfe20ff0-b434-4f25-98b2-b50b6de30fca" containerName="extract-content" Dec 04 04:59:25 crc kubenswrapper[4806]: I1204 04:59:25.943481 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfe20ff0-b434-4f25-98b2-b50b6de30fca" containerName="extract-content" Dec 04 04:59:25 crc kubenswrapper[4806]: E1204 04:59:25.943534 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfe20ff0-b434-4f25-98b2-b50b6de30fca" containerName="extract-utilities" Dec 04 04:59:25 crc kubenswrapper[4806]: I1204 04:59:25.943543 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfe20ff0-b434-4f25-98b2-b50b6de30fca" containerName="extract-utilities" Dec 04 04:59:25 crc kubenswrapper[4806]: E1204 04:59:25.943559 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfe20ff0-b434-4f25-98b2-b50b6de30fca" containerName="registry-server" Dec 04 04:59:25 crc kubenswrapper[4806]: I1204 04:59:25.943565 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfe20ff0-b434-4f25-98b2-b50b6de30fca" containerName="registry-server" Dec 04 04:59:25 crc kubenswrapper[4806]: I1204 04:59:25.943742 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfe20ff0-b434-4f25-98b2-b50b6de30fca" containerName="registry-server" Dec 04 04:59:25 crc kubenswrapper[4806]: I1204 04:59:25.945005 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:25 crc kubenswrapper[4806]: I1204 04:59:25.962987 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x8hzn"] Dec 04 04:59:26 crc kubenswrapper[4806]: I1204 04:59:26.080820 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-utilities\") pod \"certified-operators-x8hzn\" (UID: \"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6\") " pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:26 crc kubenswrapper[4806]: I1204 04:59:26.080985 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8fpd\" (UniqueName: \"kubernetes.io/projected/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-kube-api-access-n8fpd\") pod \"certified-operators-x8hzn\" (UID: \"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6\") " pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:26 crc kubenswrapper[4806]: I1204 04:59:26.081074 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-catalog-content\") pod \"certified-operators-x8hzn\" (UID: \"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6\") " pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:26 crc kubenswrapper[4806]: I1204 04:59:26.182915 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-utilities\") pod \"certified-operators-x8hzn\" (UID: \"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6\") " pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:26 crc kubenswrapper[4806]: I1204 04:59:26.183034 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8fpd\" (UniqueName: \"kubernetes.io/projected/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-kube-api-access-n8fpd\") pod \"certified-operators-x8hzn\" (UID: \"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6\") " pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:26 crc kubenswrapper[4806]: I1204 04:59:26.183112 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-catalog-content\") pod \"certified-operators-x8hzn\" (UID: \"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6\") " pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:26 crc kubenswrapper[4806]: I1204 04:59:26.183431 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-utilities\") pod \"certified-operators-x8hzn\" (UID: \"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6\") " pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:26 crc kubenswrapper[4806]: I1204 04:59:26.183454 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-catalog-content\") pod \"certified-operators-x8hzn\" (UID: \"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6\") " pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:26 crc kubenswrapper[4806]: I1204 04:59:26.207640 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8fpd\" (UniqueName: \"kubernetes.io/projected/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-kube-api-access-n8fpd\") pod \"certified-operators-x8hzn\" (UID: \"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6\") " pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:26 crc kubenswrapper[4806]: I1204 04:59:26.277537 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:26 crc kubenswrapper[4806]: I1204 04:59:26.901083 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x8hzn"] Dec 04 04:59:27 crc kubenswrapper[4806]: I1204 04:59:27.159184 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x8hzn" event={"ID":"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6","Type":"ContainerStarted","Data":"d108b3a8df0279fccd1329c718fb8a9f6970351a6c86394349989d2f0d144900"} Dec 04 04:59:27 crc kubenswrapper[4806]: I1204 04:59:27.160395 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x8hzn" event={"ID":"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6","Type":"ContainerStarted","Data":"346a781708cfe65ecf85e4b8622c271827dabed5ac784452ba8d2868075f6479"} Dec 04 04:59:27 crc kubenswrapper[4806]: I1204 04:59:27.162240 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 04:59:28 crc kubenswrapper[4806]: I1204 04:59:28.180101 4806 generic.go:334] "Generic (PLEG): container finished" podID="09b8b8cd-80f2-426e-98d2-364fd0ae4ca6" containerID="d108b3a8df0279fccd1329c718fb8a9f6970351a6c86394349989d2f0d144900" exitCode=0 Dec 04 04:59:28 crc kubenswrapper[4806]: I1204 04:59:28.180438 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x8hzn" event={"ID":"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6","Type":"ContainerDied","Data":"d108b3a8df0279fccd1329c718fb8a9f6970351a6c86394349989d2f0d144900"} Dec 04 04:59:28 crc kubenswrapper[4806]: I1204 04:59:28.180477 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x8hzn" event={"ID":"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6","Type":"ContainerStarted","Data":"f864eac6933a10ba81813a16aca6bd296f1dc045c472de2794c912318115d0cb"} Dec 04 04:59:29 crc kubenswrapper[4806]: I1204 04:59:29.204989 4806 generic.go:334] "Generic (PLEG): container finished" podID="09b8b8cd-80f2-426e-98d2-364fd0ae4ca6" containerID="f864eac6933a10ba81813a16aca6bd296f1dc045c472de2794c912318115d0cb" exitCode=0 Dec 04 04:59:29 crc kubenswrapper[4806]: I1204 04:59:29.205254 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x8hzn" event={"ID":"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6","Type":"ContainerDied","Data":"f864eac6933a10ba81813a16aca6bd296f1dc045c472de2794c912318115d0cb"} Dec 04 04:59:30 crc kubenswrapper[4806]: I1204 04:59:30.216909 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x8hzn" event={"ID":"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6","Type":"ContainerStarted","Data":"0f01a4b43fb1be03ac1fa8396ddb37755d96a38ea27486d304ad83117b6ce272"} Dec 04 04:59:30 crc kubenswrapper[4806]: I1204 04:59:30.244246 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-x8hzn" podStartSLOduration=2.74793425 podStartE2EDuration="5.244224309s" podCreationTimestamp="2025-12-04 04:59:25 +0000 UTC" firstStartedPulling="2025-12-04 04:59:27.161586488 +0000 UTC m=+3882.020099416" lastFinishedPulling="2025-12-04 04:59:29.657876547 +0000 UTC m=+3884.516389475" observedRunningTime="2025-12-04 04:59:30.233500131 +0000 UTC m=+3885.092013079" watchObservedRunningTime="2025-12-04 04:59:30.244224309 +0000 UTC m=+3885.102737237" Dec 04 04:59:36 crc kubenswrapper[4806]: I1204 04:59:36.280423 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:36 crc kubenswrapper[4806]: I1204 04:59:36.283075 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:36 crc kubenswrapper[4806]: I1204 04:59:36.354993 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:37 crc kubenswrapper[4806]: I1204 04:59:37.343319 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:37 crc kubenswrapper[4806]: I1204 04:59:37.392174 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x8hzn"] Dec 04 04:59:39 crc kubenswrapper[4806]: I1204 04:59:39.311714 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-x8hzn" podUID="09b8b8cd-80f2-426e-98d2-364fd0ae4ca6" containerName="registry-server" containerID="cri-o://0f01a4b43fb1be03ac1fa8396ddb37755d96a38ea27486d304ad83117b6ce272" gracePeriod=2 Dec 04 04:59:39 crc kubenswrapper[4806]: I1204 04:59:39.957544 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.052103 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8fpd\" (UniqueName: \"kubernetes.io/projected/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-kube-api-access-n8fpd\") pod \"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6\" (UID: \"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6\") " Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.052460 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-utilities\") pod \"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6\" (UID: \"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6\") " Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.052530 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-catalog-content\") pod \"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6\" (UID: \"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6\") " Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.054363 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-utilities" (OuterVolumeSpecName: "utilities") pod "09b8b8cd-80f2-426e-98d2-364fd0ae4ca6" (UID: "09b8b8cd-80f2-426e-98d2-364fd0ae4ca6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.273981 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.285378 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-kube-api-access-n8fpd" (OuterVolumeSpecName: "kube-api-access-n8fpd") pod "09b8b8cd-80f2-426e-98d2-364fd0ae4ca6" (UID: "09b8b8cd-80f2-426e-98d2-364fd0ae4ca6"). InnerVolumeSpecName "kube-api-access-n8fpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.352314 4806 generic.go:334] "Generic (PLEG): container finished" podID="09b8b8cd-80f2-426e-98d2-364fd0ae4ca6" containerID="0f01a4b43fb1be03ac1fa8396ddb37755d96a38ea27486d304ad83117b6ce272" exitCode=0 Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.352361 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x8hzn" event={"ID":"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6","Type":"ContainerDied","Data":"0f01a4b43fb1be03ac1fa8396ddb37755d96a38ea27486d304ad83117b6ce272"} Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.352419 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x8hzn" event={"ID":"09b8b8cd-80f2-426e-98d2-364fd0ae4ca6","Type":"ContainerDied","Data":"346a781708cfe65ecf85e4b8622c271827dabed5ac784452ba8d2868075f6479"} Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.352457 4806 scope.go:117] "RemoveContainer" containerID="0f01a4b43fb1be03ac1fa8396ddb37755d96a38ea27486d304ad83117b6ce272" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.352682 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x8hzn" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.380567 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8fpd\" (UniqueName: \"kubernetes.io/projected/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-kube-api-access-n8fpd\") on node \"crc\" DevicePath \"\"" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.411941 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09b8b8cd-80f2-426e-98d2-364fd0ae4ca6" (UID: "09b8b8cd-80f2-426e-98d2-364fd0ae4ca6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.414125 4806 scope.go:117] "RemoveContainer" containerID="f864eac6933a10ba81813a16aca6bd296f1dc045c472de2794c912318115d0cb" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.468268 4806 scope.go:117] "RemoveContainer" containerID="d108b3a8df0279fccd1329c718fb8a9f6970351a6c86394349989d2f0d144900" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.481696 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.533138 4806 scope.go:117] "RemoveContainer" containerID="0f01a4b43fb1be03ac1fa8396ddb37755d96a38ea27486d304ad83117b6ce272" Dec 04 04:59:40 crc kubenswrapper[4806]: E1204 04:59:40.536671 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f01a4b43fb1be03ac1fa8396ddb37755d96a38ea27486d304ad83117b6ce272\": container with ID starting with 0f01a4b43fb1be03ac1fa8396ddb37755d96a38ea27486d304ad83117b6ce272 not found: ID does not exist" containerID="0f01a4b43fb1be03ac1fa8396ddb37755d96a38ea27486d304ad83117b6ce272" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.536719 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f01a4b43fb1be03ac1fa8396ddb37755d96a38ea27486d304ad83117b6ce272"} err="failed to get container status \"0f01a4b43fb1be03ac1fa8396ddb37755d96a38ea27486d304ad83117b6ce272\": rpc error: code = NotFound desc = could not find container \"0f01a4b43fb1be03ac1fa8396ddb37755d96a38ea27486d304ad83117b6ce272\": container with ID starting with 0f01a4b43fb1be03ac1fa8396ddb37755d96a38ea27486d304ad83117b6ce272 not found: ID does not exist" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.536744 4806 scope.go:117] "RemoveContainer" containerID="f864eac6933a10ba81813a16aca6bd296f1dc045c472de2794c912318115d0cb" Dec 04 04:59:40 crc kubenswrapper[4806]: E1204 04:59:40.537188 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f864eac6933a10ba81813a16aca6bd296f1dc045c472de2794c912318115d0cb\": container with ID starting with f864eac6933a10ba81813a16aca6bd296f1dc045c472de2794c912318115d0cb not found: ID does not exist" containerID="f864eac6933a10ba81813a16aca6bd296f1dc045c472de2794c912318115d0cb" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.537237 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f864eac6933a10ba81813a16aca6bd296f1dc045c472de2794c912318115d0cb"} err="failed to get container status \"f864eac6933a10ba81813a16aca6bd296f1dc045c472de2794c912318115d0cb\": rpc error: code = NotFound desc = could not find container \"f864eac6933a10ba81813a16aca6bd296f1dc045c472de2794c912318115d0cb\": container with ID starting with f864eac6933a10ba81813a16aca6bd296f1dc045c472de2794c912318115d0cb not found: ID does not exist" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.537278 4806 scope.go:117] "RemoveContainer" containerID="d108b3a8df0279fccd1329c718fb8a9f6970351a6c86394349989d2f0d144900" Dec 04 04:59:40 crc kubenswrapper[4806]: E1204 04:59:40.537736 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d108b3a8df0279fccd1329c718fb8a9f6970351a6c86394349989d2f0d144900\": container with ID starting with d108b3a8df0279fccd1329c718fb8a9f6970351a6c86394349989d2f0d144900 not found: ID does not exist" containerID="d108b3a8df0279fccd1329c718fb8a9f6970351a6c86394349989d2f0d144900" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.537762 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d108b3a8df0279fccd1329c718fb8a9f6970351a6c86394349989d2f0d144900"} err="failed to get container status \"d108b3a8df0279fccd1329c718fb8a9f6970351a6c86394349989d2f0d144900\": rpc error: code = NotFound desc = could not find container \"d108b3a8df0279fccd1329c718fb8a9f6970351a6c86394349989d2f0d144900\": container with ID starting with d108b3a8df0279fccd1329c718fb8a9f6970351a6c86394349989d2f0d144900 not found: ID does not exist" Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.687896 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x8hzn"] Dec 04 04:59:40 crc kubenswrapper[4806]: I1204 04:59:40.698253 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-x8hzn"] Dec 04 04:59:41 crc kubenswrapper[4806]: I1204 04:59:41.434150 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09b8b8cd-80f2-426e-98d2-364fd0ae4ca6" path="/var/lib/kubelet/pods/09b8b8cd-80f2-426e-98d2-364fd0ae4ca6/volumes" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.192595 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84"] Dec 04 05:00:00 crc kubenswrapper[4806]: E1204 05:00:00.193688 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09b8b8cd-80f2-426e-98d2-364fd0ae4ca6" containerName="extract-utilities" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.193709 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="09b8b8cd-80f2-426e-98d2-364fd0ae4ca6" containerName="extract-utilities" Dec 04 05:00:00 crc kubenswrapper[4806]: E1204 05:00:00.193744 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09b8b8cd-80f2-426e-98d2-364fd0ae4ca6" containerName="extract-content" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.193751 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="09b8b8cd-80f2-426e-98d2-364fd0ae4ca6" containerName="extract-content" Dec 04 05:00:00 crc kubenswrapper[4806]: E1204 05:00:00.193766 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09b8b8cd-80f2-426e-98d2-364fd0ae4ca6" containerName="registry-server" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.193775 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="09b8b8cd-80f2-426e-98d2-364fd0ae4ca6" containerName="registry-server" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.196410 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="09b8b8cd-80f2-426e-98d2-364fd0ae4ca6" containerName="registry-server" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.197260 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.213879 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84"] Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.228535 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.232168 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.257547 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5df316a6-94d0-47b4-a2b5-66670f86427d-secret-volume\") pod \"collect-profiles-29413740-b8n84\" (UID: \"5df316a6-94d0-47b4-a2b5-66670f86427d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.257618 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkpw7\" (UniqueName: \"kubernetes.io/projected/5df316a6-94d0-47b4-a2b5-66670f86427d-kube-api-access-hkpw7\") pod \"collect-profiles-29413740-b8n84\" (UID: \"5df316a6-94d0-47b4-a2b5-66670f86427d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.257832 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5df316a6-94d0-47b4-a2b5-66670f86427d-config-volume\") pod \"collect-profiles-29413740-b8n84\" (UID: \"5df316a6-94d0-47b4-a2b5-66670f86427d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.359611 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5df316a6-94d0-47b4-a2b5-66670f86427d-secret-volume\") pod \"collect-profiles-29413740-b8n84\" (UID: \"5df316a6-94d0-47b4-a2b5-66670f86427d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.359669 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkpw7\" (UniqueName: \"kubernetes.io/projected/5df316a6-94d0-47b4-a2b5-66670f86427d-kube-api-access-hkpw7\") pod \"collect-profiles-29413740-b8n84\" (UID: \"5df316a6-94d0-47b4-a2b5-66670f86427d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.359785 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5df316a6-94d0-47b4-a2b5-66670f86427d-config-volume\") pod \"collect-profiles-29413740-b8n84\" (UID: \"5df316a6-94d0-47b4-a2b5-66670f86427d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.360731 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5df316a6-94d0-47b4-a2b5-66670f86427d-config-volume\") pod \"collect-profiles-29413740-b8n84\" (UID: \"5df316a6-94d0-47b4-a2b5-66670f86427d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.367877 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5df316a6-94d0-47b4-a2b5-66670f86427d-secret-volume\") pod \"collect-profiles-29413740-b8n84\" (UID: \"5df316a6-94d0-47b4-a2b5-66670f86427d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.381363 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkpw7\" (UniqueName: \"kubernetes.io/projected/5df316a6-94d0-47b4-a2b5-66670f86427d-kube-api-access-hkpw7\") pod \"collect-profiles-29413740-b8n84\" (UID: \"5df316a6-94d0-47b4-a2b5-66670f86427d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84" Dec 04 05:00:00 crc kubenswrapper[4806]: I1204 05:00:00.553898 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84" Dec 04 05:00:01 crc kubenswrapper[4806]: I1204 05:00:01.047016 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84"] Dec 04 05:00:01 crc kubenswrapper[4806]: I1204 05:00:01.547823 4806 generic.go:334] "Generic (PLEG): container finished" podID="5df316a6-94d0-47b4-a2b5-66670f86427d" containerID="978af72418ff3a000da02ad2254b9bc581932fcdbb84eb943c0e5fa52ac8080c" exitCode=0 Dec 04 05:00:01 crc kubenswrapper[4806]: I1204 05:00:01.547949 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84" event={"ID":"5df316a6-94d0-47b4-a2b5-66670f86427d","Type":"ContainerDied","Data":"978af72418ff3a000da02ad2254b9bc581932fcdbb84eb943c0e5fa52ac8080c"} Dec 04 05:00:01 crc kubenswrapper[4806]: I1204 05:00:01.548186 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84" event={"ID":"5df316a6-94d0-47b4-a2b5-66670f86427d","Type":"ContainerStarted","Data":"3652778269df590d29bbe0be9ca3c4cb11baf6b59b0dd33a98610fde7971b1b4"} Dec 04 05:00:02 crc kubenswrapper[4806]: I1204 05:00:02.286415 4806 scope.go:117] "RemoveContainer" containerID="31c8c52f613d030396e3232137da6cde4871ca350b3112369b28197792bbe2e4" Dec 04 05:00:02 crc kubenswrapper[4806]: I1204 05:00:02.333098 4806 scope.go:117] "RemoveContainer" containerID="cb9c69372b7785e69feebf6ed4a8e538bf71951d82cb75fb4978294079d4bcbc" Dec 04 05:00:02 crc kubenswrapper[4806]: I1204 05:00:02.354380 4806 scope.go:117] "RemoveContainer" containerID="912294280f99d80041d2c035c6c28f9a3b9aae67d5742d8820de3351db533217" Dec 04 05:00:03 crc kubenswrapper[4806]: I1204 05:00:03.010054 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84" Dec 04 05:00:03 crc kubenswrapper[4806]: I1204 05:00:03.124903 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5df316a6-94d0-47b4-a2b5-66670f86427d-config-volume\") pod \"5df316a6-94d0-47b4-a2b5-66670f86427d\" (UID: \"5df316a6-94d0-47b4-a2b5-66670f86427d\") " Dec 04 05:00:03 crc kubenswrapper[4806]: I1204 05:00:03.125045 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkpw7\" (UniqueName: \"kubernetes.io/projected/5df316a6-94d0-47b4-a2b5-66670f86427d-kube-api-access-hkpw7\") pod \"5df316a6-94d0-47b4-a2b5-66670f86427d\" (UID: \"5df316a6-94d0-47b4-a2b5-66670f86427d\") " Dec 04 05:00:03 crc kubenswrapper[4806]: I1204 05:00:03.125264 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5df316a6-94d0-47b4-a2b5-66670f86427d-secret-volume\") pod \"5df316a6-94d0-47b4-a2b5-66670f86427d\" (UID: \"5df316a6-94d0-47b4-a2b5-66670f86427d\") " Dec 04 05:00:03 crc kubenswrapper[4806]: I1204 05:00:03.125826 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5df316a6-94d0-47b4-a2b5-66670f86427d-config-volume" (OuterVolumeSpecName: "config-volume") pod "5df316a6-94d0-47b4-a2b5-66670f86427d" (UID: "5df316a6-94d0-47b4-a2b5-66670f86427d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 05:00:03 crc kubenswrapper[4806]: I1204 05:00:03.126689 4806 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5df316a6-94d0-47b4-a2b5-66670f86427d-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 05:00:03 crc kubenswrapper[4806]: I1204 05:00:03.133188 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5df316a6-94d0-47b4-a2b5-66670f86427d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5df316a6-94d0-47b4-a2b5-66670f86427d" (UID: "5df316a6-94d0-47b4-a2b5-66670f86427d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 05:00:03 crc kubenswrapper[4806]: I1204 05:00:03.133306 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5df316a6-94d0-47b4-a2b5-66670f86427d-kube-api-access-hkpw7" (OuterVolumeSpecName: "kube-api-access-hkpw7") pod "5df316a6-94d0-47b4-a2b5-66670f86427d" (UID: "5df316a6-94d0-47b4-a2b5-66670f86427d"). InnerVolumeSpecName "kube-api-access-hkpw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:00:03 crc kubenswrapper[4806]: I1204 05:00:03.228880 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkpw7\" (UniqueName: \"kubernetes.io/projected/5df316a6-94d0-47b4-a2b5-66670f86427d-kube-api-access-hkpw7\") on node \"crc\" DevicePath \"\"" Dec 04 05:00:03 crc kubenswrapper[4806]: I1204 05:00:03.228977 4806 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5df316a6-94d0-47b4-a2b5-66670f86427d-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 05:00:03 crc kubenswrapper[4806]: I1204 05:00:03.570760 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84" event={"ID":"5df316a6-94d0-47b4-a2b5-66670f86427d","Type":"ContainerDied","Data":"3652778269df590d29bbe0be9ca3c4cb11baf6b59b0dd33a98610fde7971b1b4"} Dec 04 05:00:03 crc kubenswrapper[4806]: I1204 05:00:03.570809 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3652778269df590d29bbe0be9ca3c4cb11baf6b59b0dd33a98610fde7971b1b4" Dec 04 05:00:03 crc kubenswrapper[4806]: I1204 05:00:03.570870 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413740-b8n84" Dec 04 05:00:04 crc kubenswrapper[4806]: I1204 05:00:04.110251 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq"] Dec 04 05:00:04 crc kubenswrapper[4806]: I1204 05:00:04.118682 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413695-thgjq"] Dec 04 05:00:05 crc kubenswrapper[4806]: I1204 05:00:05.439161 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d855714a-f641-44bc-9b95-ebda879968b0" path="/var/lib/kubelet/pods/d855714a-f641-44bc-9b95-ebda879968b0/volumes" Dec 04 05:00:30 crc kubenswrapper[4806]: I1204 05:00:30.591275 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v5zw9"] Dec 04 05:00:30 crc kubenswrapper[4806]: E1204 05:00:30.602361 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5df316a6-94d0-47b4-a2b5-66670f86427d" containerName="collect-profiles" Dec 04 05:00:30 crc kubenswrapper[4806]: I1204 05:00:30.602391 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="5df316a6-94d0-47b4-a2b5-66670f86427d" containerName="collect-profiles" Dec 04 05:00:30 crc kubenswrapper[4806]: I1204 05:00:30.602655 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="5df316a6-94d0-47b4-a2b5-66670f86427d" containerName="collect-profiles" Dec 04 05:00:30 crc kubenswrapper[4806]: I1204 05:00:30.604754 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:30 crc kubenswrapper[4806]: I1204 05:00:30.612507 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v5zw9"] Dec 04 05:00:30 crc kubenswrapper[4806]: I1204 05:00:30.773318 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd9b102c-bd71-41f4-94d7-c969fdc977b6-catalog-content\") pod \"redhat-operators-v5zw9\" (UID: \"bd9b102c-bd71-41f4-94d7-c969fdc977b6\") " pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:30 crc kubenswrapper[4806]: I1204 05:00:30.773354 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvjdh\" (UniqueName: \"kubernetes.io/projected/bd9b102c-bd71-41f4-94d7-c969fdc977b6-kube-api-access-zvjdh\") pod \"redhat-operators-v5zw9\" (UID: \"bd9b102c-bd71-41f4-94d7-c969fdc977b6\") " pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:30 crc kubenswrapper[4806]: I1204 05:00:30.773389 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd9b102c-bd71-41f4-94d7-c969fdc977b6-utilities\") pod \"redhat-operators-v5zw9\" (UID: \"bd9b102c-bd71-41f4-94d7-c969fdc977b6\") " pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:30 crc kubenswrapper[4806]: I1204 05:00:30.875680 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd9b102c-bd71-41f4-94d7-c969fdc977b6-catalog-content\") pod \"redhat-operators-v5zw9\" (UID: \"bd9b102c-bd71-41f4-94d7-c969fdc977b6\") " pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:30 crc kubenswrapper[4806]: I1204 05:00:30.875754 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvjdh\" (UniqueName: \"kubernetes.io/projected/bd9b102c-bd71-41f4-94d7-c969fdc977b6-kube-api-access-zvjdh\") pod \"redhat-operators-v5zw9\" (UID: \"bd9b102c-bd71-41f4-94d7-c969fdc977b6\") " pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:30 crc kubenswrapper[4806]: I1204 05:00:30.875789 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd9b102c-bd71-41f4-94d7-c969fdc977b6-utilities\") pod \"redhat-operators-v5zw9\" (UID: \"bd9b102c-bd71-41f4-94d7-c969fdc977b6\") " pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:30 crc kubenswrapper[4806]: I1204 05:00:30.876234 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd9b102c-bd71-41f4-94d7-c969fdc977b6-utilities\") pod \"redhat-operators-v5zw9\" (UID: \"bd9b102c-bd71-41f4-94d7-c969fdc977b6\") " pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:30 crc kubenswrapper[4806]: I1204 05:00:30.876268 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd9b102c-bd71-41f4-94d7-c969fdc977b6-catalog-content\") pod \"redhat-operators-v5zw9\" (UID: \"bd9b102c-bd71-41f4-94d7-c969fdc977b6\") " pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:31 crc kubenswrapper[4806]: I1204 05:00:31.153778 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvjdh\" (UniqueName: \"kubernetes.io/projected/bd9b102c-bd71-41f4-94d7-c969fdc977b6-kube-api-access-zvjdh\") pod \"redhat-operators-v5zw9\" (UID: \"bd9b102c-bd71-41f4-94d7-c969fdc977b6\") " pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:31 crc kubenswrapper[4806]: I1204 05:00:31.245918 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:31 crc kubenswrapper[4806]: I1204 05:00:31.880202 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v5zw9"] Dec 04 05:00:32 crc kubenswrapper[4806]: I1204 05:00:32.852168 4806 generic.go:334] "Generic (PLEG): container finished" podID="bd9b102c-bd71-41f4-94d7-c969fdc977b6" containerID="a7e5922815c0e542918cead39f1c0066860bb44f070829689c6e943807f0e4aa" exitCode=0 Dec 04 05:00:32 crc kubenswrapper[4806]: I1204 05:00:32.853170 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5zw9" event={"ID":"bd9b102c-bd71-41f4-94d7-c969fdc977b6","Type":"ContainerDied","Data":"a7e5922815c0e542918cead39f1c0066860bb44f070829689c6e943807f0e4aa"} Dec 04 05:00:32 crc kubenswrapper[4806]: I1204 05:00:32.853250 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5zw9" event={"ID":"bd9b102c-bd71-41f4-94d7-c969fdc977b6","Type":"ContainerStarted","Data":"11688e135aff2e70c73dddfdefcd5fc2e246cd13a416cb3c2f4c6d1b8c1ad3e2"} Dec 04 05:00:33 crc kubenswrapper[4806]: I1204 05:00:33.867042 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5zw9" event={"ID":"bd9b102c-bd71-41f4-94d7-c969fdc977b6","Type":"ContainerStarted","Data":"40b8f6a25361723df215ac8744f894833de2ef455a4e0092e5a30e05c5f574f6"} Dec 04 05:00:36 crc kubenswrapper[4806]: I1204 05:00:36.891525 4806 generic.go:334] "Generic (PLEG): container finished" podID="bd9b102c-bd71-41f4-94d7-c969fdc977b6" containerID="40b8f6a25361723df215ac8744f894833de2ef455a4e0092e5a30e05c5f574f6" exitCode=0 Dec 04 05:00:36 crc kubenswrapper[4806]: I1204 05:00:36.891778 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5zw9" event={"ID":"bd9b102c-bd71-41f4-94d7-c969fdc977b6","Type":"ContainerDied","Data":"40b8f6a25361723df215ac8744f894833de2ef455a4e0092e5a30e05c5f574f6"} Dec 04 05:00:37 crc kubenswrapper[4806]: I1204 05:00:37.906221 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5zw9" event={"ID":"bd9b102c-bd71-41f4-94d7-c969fdc977b6","Type":"ContainerStarted","Data":"b13b0e797438424d49216e345587e2577419cda011e77de15c92b3470ac97d5e"} Dec 04 05:00:37 crc kubenswrapper[4806]: I1204 05:00:37.932159 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v5zw9" podStartSLOduration=3.4663956430000002 podStartE2EDuration="7.932139228s" podCreationTimestamp="2025-12-04 05:00:30 +0000 UTC" firstStartedPulling="2025-12-04 05:00:32.856605442 +0000 UTC m=+3947.715118370" lastFinishedPulling="2025-12-04 05:00:37.322349027 +0000 UTC m=+3952.180861955" observedRunningTime="2025-12-04 05:00:37.924338412 +0000 UTC m=+3952.782851340" watchObservedRunningTime="2025-12-04 05:00:37.932139228 +0000 UTC m=+3952.790652156" Dec 04 05:00:41 crc kubenswrapper[4806]: I1204 05:00:41.246499 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:41 crc kubenswrapper[4806]: I1204 05:00:41.247301 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:42 crc kubenswrapper[4806]: I1204 05:00:42.697375 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v5zw9" podUID="bd9b102c-bd71-41f4-94d7-c969fdc977b6" containerName="registry-server" probeResult="failure" output=< Dec 04 05:00:42 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Dec 04 05:00:42 crc kubenswrapper[4806]: > Dec 04 05:00:51 crc kubenswrapper[4806]: I1204 05:00:51.504551 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:51 crc kubenswrapper[4806]: I1204 05:00:51.572446 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:51 crc kubenswrapper[4806]: I1204 05:00:51.761676 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v5zw9"] Dec 04 05:00:53 crc kubenswrapper[4806]: I1204 05:00:53.038839 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v5zw9" podUID="bd9b102c-bd71-41f4-94d7-c969fdc977b6" containerName="registry-server" containerID="cri-o://b13b0e797438424d49216e345587e2577419cda011e77de15c92b3470ac97d5e" gracePeriod=2 Dec 04 05:00:53 crc kubenswrapper[4806]: I1204 05:00:53.967734 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.055797 4806 generic.go:334] "Generic (PLEG): container finished" podID="bd9b102c-bd71-41f4-94d7-c969fdc977b6" containerID="b13b0e797438424d49216e345587e2577419cda011e77de15c92b3470ac97d5e" exitCode=0 Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.055844 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5zw9" event={"ID":"bd9b102c-bd71-41f4-94d7-c969fdc977b6","Type":"ContainerDied","Data":"b13b0e797438424d49216e345587e2577419cda011e77de15c92b3470ac97d5e"} Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.055873 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5zw9" event={"ID":"bd9b102c-bd71-41f4-94d7-c969fdc977b6","Type":"ContainerDied","Data":"11688e135aff2e70c73dddfdefcd5fc2e246cd13a416cb3c2f4c6d1b8c1ad3e2"} Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.055893 4806 scope.go:117] "RemoveContainer" containerID="b13b0e797438424d49216e345587e2577419cda011e77de15c92b3470ac97d5e" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.055976 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v5zw9" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.099659 4806 scope.go:117] "RemoveContainer" containerID="40b8f6a25361723df215ac8744f894833de2ef455a4e0092e5a30e05c5f574f6" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.132237 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvjdh\" (UniqueName: \"kubernetes.io/projected/bd9b102c-bd71-41f4-94d7-c969fdc977b6-kube-api-access-zvjdh\") pod \"bd9b102c-bd71-41f4-94d7-c969fdc977b6\" (UID: \"bd9b102c-bd71-41f4-94d7-c969fdc977b6\") " Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.132324 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd9b102c-bd71-41f4-94d7-c969fdc977b6-catalog-content\") pod \"bd9b102c-bd71-41f4-94d7-c969fdc977b6\" (UID: \"bd9b102c-bd71-41f4-94d7-c969fdc977b6\") " Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.132378 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd9b102c-bd71-41f4-94d7-c969fdc977b6-utilities\") pod \"bd9b102c-bd71-41f4-94d7-c969fdc977b6\" (UID: \"bd9b102c-bd71-41f4-94d7-c969fdc977b6\") " Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.133919 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd9b102c-bd71-41f4-94d7-c969fdc977b6-utilities" (OuterVolumeSpecName: "utilities") pod "bd9b102c-bd71-41f4-94d7-c969fdc977b6" (UID: "bd9b102c-bd71-41f4-94d7-c969fdc977b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.163751 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd9b102c-bd71-41f4-94d7-c969fdc977b6-kube-api-access-zvjdh" (OuterVolumeSpecName: "kube-api-access-zvjdh") pod "bd9b102c-bd71-41f4-94d7-c969fdc977b6" (UID: "bd9b102c-bd71-41f4-94d7-c969fdc977b6"). InnerVolumeSpecName "kube-api-access-zvjdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.212659 4806 scope.go:117] "RemoveContainer" containerID="a7e5922815c0e542918cead39f1c0066860bb44f070829689c6e943807f0e4aa" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.236584 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvjdh\" (UniqueName: \"kubernetes.io/projected/bd9b102c-bd71-41f4-94d7-c969fdc977b6-kube-api-access-zvjdh\") on node \"crc\" DevicePath \"\"" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.236614 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd9b102c-bd71-41f4-94d7-c969fdc977b6-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.278765 4806 scope.go:117] "RemoveContainer" containerID="b13b0e797438424d49216e345587e2577419cda011e77de15c92b3470ac97d5e" Dec 04 05:00:54 crc kubenswrapper[4806]: E1204 05:00:54.279126 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b13b0e797438424d49216e345587e2577419cda011e77de15c92b3470ac97d5e\": container with ID starting with b13b0e797438424d49216e345587e2577419cda011e77de15c92b3470ac97d5e not found: ID does not exist" containerID="b13b0e797438424d49216e345587e2577419cda011e77de15c92b3470ac97d5e" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.279157 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b13b0e797438424d49216e345587e2577419cda011e77de15c92b3470ac97d5e"} err="failed to get container status \"b13b0e797438424d49216e345587e2577419cda011e77de15c92b3470ac97d5e\": rpc error: code = NotFound desc = could not find container \"b13b0e797438424d49216e345587e2577419cda011e77de15c92b3470ac97d5e\": container with ID starting with b13b0e797438424d49216e345587e2577419cda011e77de15c92b3470ac97d5e not found: ID does not exist" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.279201 4806 scope.go:117] "RemoveContainer" containerID="40b8f6a25361723df215ac8744f894833de2ef455a4e0092e5a30e05c5f574f6" Dec 04 05:00:54 crc kubenswrapper[4806]: E1204 05:00:54.280878 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40b8f6a25361723df215ac8744f894833de2ef455a4e0092e5a30e05c5f574f6\": container with ID starting with 40b8f6a25361723df215ac8744f894833de2ef455a4e0092e5a30e05c5f574f6 not found: ID does not exist" containerID="40b8f6a25361723df215ac8744f894833de2ef455a4e0092e5a30e05c5f574f6" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.280947 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40b8f6a25361723df215ac8744f894833de2ef455a4e0092e5a30e05c5f574f6"} err="failed to get container status \"40b8f6a25361723df215ac8744f894833de2ef455a4e0092e5a30e05c5f574f6\": rpc error: code = NotFound desc = could not find container \"40b8f6a25361723df215ac8744f894833de2ef455a4e0092e5a30e05c5f574f6\": container with ID starting with 40b8f6a25361723df215ac8744f894833de2ef455a4e0092e5a30e05c5f574f6 not found: ID does not exist" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.280979 4806 scope.go:117] "RemoveContainer" containerID="a7e5922815c0e542918cead39f1c0066860bb44f070829689c6e943807f0e4aa" Dec 04 05:00:54 crc kubenswrapper[4806]: E1204 05:00:54.281340 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7e5922815c0e542918cead39f1c0066860bb44f070829689c6e943807f0e4aa\": container with ID starting with a7e5922815c0e542918cead39f1c0066860bb44f070829689c6e943807f0e4aa not found: ID does not exist" containerID="a7e5922815c0e542918cead39f1c0066860bb44f070829689c6e943807f0e4aa" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.281382 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7e5922815c0e542918cead39f1c0066860bb44f070829689c6e943807f0e4aa"} err="failed to get container status \"a7e5922815c0e542918cead39f1c0066860bb44f070829689c6e943807f0e4aa\": rpc error: code = NotFound desc = could not find container \"a7e5922815c0e542918cead39f1c0066860bb44f070829689c6e943807f0e4aa\": container with ID starting with a7e5922815c0e542918cead39f1c0066860bb44f070829689c6e943807f0e4aa not found: ID does not exist" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.306694 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd9b102c-bd71-41f4-94d7-c969fdc977b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bd9b102c-bd71-41f4-94d7-c969fdc977b6" (UID: "bd9b102c-bd71-41f4-94d7-c969fdc977b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.338620 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd9b102c-bd71-41f4-94d7-c969fdc977b6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.425625 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v5zw9"] Dec 04 05:00:54 crc kubenswrapper[4806]: I1204 05:00:54.434635 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v5zw9"] Dec 04 05:00:55 crc kubenswrapper[4806]: I1204 05:00:55.435361 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd9b102c-bd71-41f4-94d7-c969fdc977b6" path="/var/lib/kubelet/pods/bd9b102c-bd71-41f4-94d7-c969fdc977b6/volumes" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.158891 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29413741-vfktv"] Dec 04 05:01:00 crc kubenswrapper[4806]: E1204 05:01:00.160074 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd9b102c-bd71-41f4-94d7-c969fdc977b6" containerName="extract-content" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.160096 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd9b102c-bd71-41f4-94d7-c969fdc977b6" containerName="extract-content" Dec 04 05:01:00 crc kubenswrapper[4806]: E1204 05:01:00.160120 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd9b102c-bd71-41f4-94d7-c969fdc977b6" containerName="extract-utilities" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.160128 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd9b102c-bd71-41f4-94d7-c969fdc977b6" containerName="extract-utilities" Dec 04 05:01:00 crc kubenswrapper[4806]: E1204 05:01:00.160139 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd9b102c-bd71-41f4-94d7-c969fdc977b6" containerName="registry-server" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.160146 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd9b102c-bd71-41f4-94d7-c969fdc977b6" containerName="registry-server" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.160388 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd9b102c-bd71-41f4-94d7-c969fdc977b6" containerName="registry-server" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.161457 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413741-vfktv" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.171829 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29413741-vfktv"] Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.289174 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7rng\" (UniqueName: \"kubernetes.io/projected/106a7dfc-57ef-4177-821d-f77a0329ac90-kube-api-access-z7rng\") pod \"keystone-cron-29413741-vfktv\" (UID: \"106a7dfc-57ef-4177-821d-f77a0329ac90\") " pod="openstack/keystone-cron-29413741-vfktv" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.289266 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-combined-ca-bundle\") pod \"keystone-cron-29413741-vfktv\" (UID: \"106a7dfc-57ef-4177-821d-f77a0329ac90\") " pod="openstack/keystone-cron-29413741-vfktv" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.289283 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-config-data\") pod \"keystone-cron-29413741-vfktv\" (UID: \"106a7dfc-57ef-4177-821d-f77a0329ac90\") " pod="openstack/keystone-cron-29413741-vfktv" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.289341 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-fernet-keys\") pod \"keystone-cron-29413741-vfktv\" (UID: \"106a7dfc-57ef-4177-821d-f77a0329ac90\") " pod="openstack/keystone-cron-29413741-vfktv" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.391604 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7rng\" (UniqueName: \"kubernetes.io/projected/106a7dfc-57ef-4177-821d-f77a0329ac90-kube-api-access-z7rng\") pod \"keystone-cron-29413741-vfktv\" (UID: \"106a7dfc-57ef-4177-821d-f77a0329ac90\") " pod="openstack/keystone-cron-29413741-vfktv" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.391691 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-config-data\") pod \"keystone-cron-29413741-vfktv\" (UID: \"106a7dfc-57ef-4177-821d-f77a0329ac90\") " pod="openstack/keystone-cron-29413741-vfktv" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.391719 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-combined-ca-bundle\") pod \"keystone-cron-29413741-vfktv\" (UID: \"106a7dfc-57ef-4177-821d-f77a0329ac90\") " pod="openstack/keystone-cron-29413741-vfktv" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.391753 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-fernet-keys\") pod \"keystone-cron-29413741-vfktv\" (UID: \"106a7dfc-57ef-4177-821d-f77a0329ac90\") " pod="openstack/keystone-cron-29413741-vfktv" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.399278 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-config-data\") pod \"keystone-cron-29413741-vfktv\" (UID: \"106a7dfc-57ef-4177-821d-f77a0329ac90\") " pod="openstack/keystone-cron-29413741-vfktv" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.399317 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-combined-ca-bundle\") pod \"keystone-cron-29413741-vfktv\" (UID: \"106a7dfc-57ef-4177-821d-f77a0329ac90\") " pod="openstack/keystone-cron-29413741-vfktv" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.414966 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-fernet-keys\") pod \"keystone-cron-29413741-vfktv\" (UID: \"106a7dfc-57ef-4177-821d-f77a0329ac90\") " pod="openstack/keystone-cron-29413741-vfktv" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.416227 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7rng\" (UniqueName: \"kubernetes.io/projected/106a7dfc-57ef-4177-821d-f77a0329ac90-kube-api-access-z7rng\") pod \"keystone-cron-29413741-vfktv\" (UID: \"106a7dfc-57ef-4177-821d-f77a0329ac90\") " pod="openstack/keystone-cron-29413741-vfktv" Dec 04 05:01:00 crc kubenswrapper[4806]: I1204 05:01:00.494459 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413741-vfktv" Dec 04 05:01:01 crc kubenswrapper[4806]: I1204 05:01:01.009716 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29413741-vfktv"] Dec 04 05:01:01 crc kubenswrapper[4806]: I1204 05:01:01.125481 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413741-vfktv" event={"ID":"106a7dfc-57ef-4177-821d-f77a0329ac90","Type":"ContainerStarted","Data":"fc026acb73b7338248a04c42865675ea9d303a0e7547af461266df97b3a894c0"} Dec 04 05:01:02 crc kubenswrapper[4806]: I1204 05:01:02.140305 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413741-vfktv" event={"ID":"106a7dfc-57ef-4177-821d-f77a0329ac90","Type":"ContainerStarted","Data":"0a421597f084fe7653b2c46dff9315605221ea30588652f23488bd1b407d015a"} Dec 04 05:01:02 crc kubenswrapper[4806]: I1204 05:01:02.159732 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29413741-vfktv" podStartSLOduration=2.159710719 podStartE2EDuration="2.159710719s" podCreationTimestamp="2025-12-04 05:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 05:01:02.159457882 +0000 UTC m=+3977.017970810" watchObservedRunningTime="2025-12-04 05:01:02.159710719 +0000 UTC m=+3977.018223647" Dec 04 05:01:02 crc kubenswrapper[4806]: I1204 05:01:02.425902 4806 scope.go:117] "RemoveContainer" containerID="581a1e4deb008f3479806b535a070be534b5d91c291b02caf9d8ede622d3103b" Dec 04 05:01:05 crc kubenswrapper[4806]: I1204 05:01:05.181706 4806 generic.go:334] "Generic (PLEG): container finished" podID="106a7dfc-57ef-4177-821d-f77a0329ac90" containerID="0a421597f084fe7653b2c46dff9315605221ea30588652f23488bd1b407d015a" exitCode=0 Dec 04 05:01:05 crc kubenswrapper[4806]: I1204 05:01:05.181846 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413741-vfktv" event={"ID":"106a7dfc-57ef-4177-821d-f77a0329ac90","Type":"ContainerDied","Data":"0a421597f084fe7653b2c46dff9315605221ea30588652f23488bd1b407d015a"} Dec 04 05:01:06 crc kubenswrapper[4806]: I1204 05:01:06.725667 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413741-vfktv" Dec 04 05:01:06 crc kubenswrapper[4806]: I1204 05:01:06.845364 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7rng\" (UniqueName: \"kubernetes.io/projected/106a7dfc-57ef-4177-821d-f77a0329ac90-kube-api-access-z7rng\") pod \"106a7dfc-57ef-4177-821d-f77a0329ac90\" (UID: \"106a7dfc-57ef-4177-821d-f77a0329ac90\") " Dec 04 05:01:06 crc kubenswrapper[4806]: I1204 05:01:06.845732 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-fernet-keys\") pod \"106a7dfc-57ef-4177-821d-f77a0329ac90\" (UID: \"106a7dfc-57ef-4177-821d-f77a0329ac90\") " Dec 04 05:01:06 crc kubenswrapper[4806]: I1204 05:01:06.845853 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-combined-ca-bundle\") pod \"106a7dfc-57ef-4177-821d-f77a0329ac90\" (UID: \"106a7dfc-57ef-4177-821d-f77a0329ac90\") " Dec 04 05:01:06 crc kubenswrapper[4806]: I1204 05:01:06.845998 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-config-data\") pod \"106a7dfc-57ef-4177-821d-f77a0329ac90\" (UID: \"106a7dfc-57ef-4177-821d-f77a0329ac90\") " Dec 04 05:01:06 crc kubenswrapper[4806]: I1204 05:01:06.853321 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/106a7dfc-57ef-4177-821d-f77a0329ac90-kube-api-access-z7rng" (OuterVolumeSpecName: "kube-api-access-z7rng") pod "106a7dfc-57ef-4177-821d-f77a0329ac90" (UID: "106a7dfc-57ef-4177-821d-f77a0329ac90"). InnerVolumeSpecName "kube-api-access-z7rng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:01:06 crc kubenswrapper[4806]: I1204 05:01:06.872835 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "106a7dfc-57ef-4177-821d-f77a0329ac90" (UID: "106a7dfc-57ef-4177-821d-f77a0329ac90"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 05:01:06 crc kubenswrapper[4806]: I1204 05:01:06.877972 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "106a7dfc-57ef-4177-821d-f77a0329ac90" (UID: "106a7dfc-57ef-4177-821d-f77a0329ac90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 05:01:06 crc kubenswrapper[4806]: I1204 05:01:06.905876 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-config-data" (OuterVolumeSpecName: "config-data") pod "106a7dfc-57ef-4177-821d-f77a0329ac90" (UID: "106a7dfc-57ef-4177-821d-f77a0329ac90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 05:01:06 crc kubenswrapper[4806]: I1204 05:01:06.947606 4806 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 05:01:06 crc kubenswrapper[4806]: I1204 05:01:06.947653 4806 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 05:01:06 crc kubenswrapper[4806]: I1204 05:01:06.947668 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/106a7dfc-57ef-4177-821d-f77a0329ac90-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 05:01:06 crc kubenswrapper[4806]: I1204 05:01:06.947853 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7rng\" (UniqueName: \"kubernetes.io/projected/106a7dfc-57ef-4177-821d-f77a0329ac90-kube-api-access-z7rng\") on node \"crc\" DevicePath \"\"" Dec 04 05:01:07 crc kubenswrapper[4806]: I1204 05:01:07.203822 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413741-vfktv" event={"ID":"106a7dfc-57ef-4177-821d-f77a0329ac90","Type":"ContainerDied","Data":"fc026acb73b7338248a04c42865675ea9d303a0e7547af461266df97b3a894c0"} Dec 04 05:01:07 crc kubenswrapper[4806]: I1204 05:01:07.203862 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc026acb73b7338248a04c42865675ea9d303a0e7547af461266df97b3a894c0" Dec 04 05:01:07 crc kubenswrapper[4806]: I1204 05:01:07.203912 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413741-vfktv" Dec 04 05:01:27 crc kubenswrapper[4806]: I1204 05:01:27.046809 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:01:27 crc kubenswrapper[4806]: I1204 05:01:27.047414 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:01:57 crc kubenswrapper[4806]: I1204 05:01:57.047569 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:01:57 crc kubenswrapper[4806]: I1204 05:01:57.048196 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:02:27 crc kubenswrapper[4806]: I1204 05:02:27.047427 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:02:27 crc kubenswrapper[4806]: I1204 05:02:27.048088 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:02:27 crc kubenswrapper[4806]: I1204 05:02:27.048156 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 05:02:27 crc kubenswrapper[4806]: I1204 05:02:27.049284 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b7a929db250bf8fd36191354c0529747a23a7c22591c6919b0c88de706949742"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 05:02:27 crc kubenswrapper[4806]: I1204 05:02:27.049363 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://b7a929db250bf8fd36191354c0529747a23a7c22591c6919b0c88de706949742" gracePeriod=600 Dec 04 05:02:27 crc kubenswrapper[4806]: I1204 05:02:27.287854 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="b7a929db250bf8fd36191354c0529747a23a7c22591c6919b0c88de706949742" exitCode=0 Dec 04 05:02:27 crc kubenswrapper[4806]: I1204 05:02:27.287956 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"b7a929db250bf8fd36191354c0529747a23a7c22591c6919b0c88de706949742"} Dec 04 05:02:27 crc kubenswrapper[4806]: I1204 05:02:27.288298 4806 scope.go:117] "RemoveContainer" containerID="b1525387f2fa0cb2a01e32f355f0edc0024beb6f6a2789b69060c321eb35dd47" Dec 04 05:02:28 crc kubenswrapper[4806]: I1204 05:02:28.298753 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744"} Dec 04 05:04:00 crc kubenswrapper[4806]: I1204 05:04:00.513009 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p55dr"] Dec 04 05:04:00 crc kubenswrapper[4806]: E1204 05:04:00.514056 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="106a7dfc-57ef-4177-821d-f77a0329ac90" containerName="keystone-cron" Dec 04 05:04:00 crc kubenswrapper[4806]: I1204 05:04:00.514074 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="106a7dfc-57ef-4177-821d-f77a0329ac90" containerName="keystone-cron" Dec 04 05:04:00 crc kubenswrapper[4806]: I1204 05:04:00.514342 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="106a7dfc-57ef-4177-821d-f77a0329ac90" containerName="keystone-cron" Dec 04 05:04:00 crc kubenswrapper[4806]: I1204 05:04:00.515784 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:00 crc kubenswrapper[4806]: I1204 05:04:00.540506 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p55dr"] Dec 04 05:04:00 crc kubenswrapper[4806]: I1204 05:04:00.644505 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed902041-2d66-4776-bd9e-5fc9f4f55048-catalog-content\") pod \"community-operators-p55dr\" (UID: \"ed902041-2d66-4776-bd9e-5fc9f4f55048\") " pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:00 crc kubenswrapper[4806]: I1204 05:04:00.644629 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed902041-2d66-4776-bd9e-5fc9f4f55048-utilities\") pod \"community-operators-p55dr\" (UID: \"ed902041-2d66-4776-bd9e-5fc9f4f55048\") " pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:00 crc kubenswrapper[4806]: I1204 05:04:00.644753 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25bsb\" (UniqueName: \"kubernetes.io/projected/ed902041-2d66-4776-bd9e-5fc9f4f55048-kube-api-access-25bsb\") pod \"community-operators-p55dr\" (UID: \"ed902041-2d66-4776-bd9e-5fc9f4f55048\") " pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:00 crc kubenswrapper[4806]: I1204 05:04:00.746834 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25bsb\" (UniqueName: \"kubernetes.io/projected/ed902041-2d66-4776-bd9e-5fc9f4f55048-kube-api-access-25bsb\") pod \"community-operators-p55dr\" (UID: \"ed902041-2d66-4776-bd9e-5fc9f4f55048\") " pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:00 crc kubenswrapper[4806]: I1204 05:04:00.746972 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed902041-2d66-4776-bd9e-5fc9f4f55048-catalog-content\") pod \"community-operators-p55dr\" (UID: \"ed902041-2d66-4776-bd9e-5fc9f4f55048\") " pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:00 crc kubenswrapper[4806]: I1204 05:04:00.747056 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed902041-2d66-4776-bd9e-5fc9f4f55048-utilities\") pod \"community-operators-p55dr\" (UID: \"ed902041-2d66-4776-bd9e-5fc9f4f55048\") " pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:00 crc kubenswrapper[4806]: I1204 05:04:00.747907 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed902041-2d66-4776-bd9e-5fc9f4f55048-catalog-content\") pod \"community-operators-p55dr\" (UID: \"ed902041-2d66-4776-bd9e-5fc9f4f55048\") " pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:00 crc kubenswrapper[4806]: I1204 05:04:00.748003 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed902041-2d66-4776-bd9e-5fc9f4f55048-utilities\") pod \"community-operators-p55dr\" (UID: \"ed902041-2d66-4776-bd9e-5fc9f4f55048\") " pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:00 crc kubenswrapper[4806]: I1204 05:04:00.782806 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25bsb\" (UniqueName: \"kubernetes.io/projected/ed902041-2d66-4776-bd9e-5fc9f4f55048-kube-api-access-25bsb\") pod \"community-operators-p55dr\" (UID: \"ed902041-2d66-4776-bd9e-5fc9f4f55048\") " pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:00 crc kubenswrapper[4806]: I1204 05:04:00.878685 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:01 crc kubenswrapper[4806]: I1204 05:04:01.474750 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p55dr"] Dec 04 05:04:02 crc kubenswrapper[4806]: I1204 05:04:02.159527 4806 generic.go:334] "Generic (PLEG): container finished" podID="ed902041-2d66-4776-bd9e-5fc9f4f55048" containerID="8df97461a0ad5ff6e070597601a66fca5f5c6078604701516ab6c7be9c9efdb6" exitCode=0 Dec 04 05:04:02 crc kubenswrapper[4806]: I1204 05:04:02.159649 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p55dr" event={"ID":"ed902041-2d66-4776-bd9e-5fc9f4f55048","Type":"ContainerDied","Data":"8df97461a0ad5ff6e070597601a66fca5f5c6078604701516ab6c7be9c9efdb6"} Dec 04 05:04:02 crc kubenswrapper[4806]: I1204 05:04:02.160307 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p55dr" event={"ID":"ed902041-2d66-4776-bd9e-5fc9f4f55048","Type":"ContainerStarted","Data":"702237eeefe301a21f27213c7f3f9dec607949eab301d8d67c6dad6eedfa3df4"} Dec 04 05:04:03 crc kubenswrapper[4806]: I1204 05:04:03.171567 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p55dr" event={"ID":"ed902041-2d66-4776-bd9e-5fc9f4f55048","Type":"ContainerStarted","Data":"75771ff539642ab63dc40c422e4057ba21b2dbf32411a7a122ca15d642bb2c07"} Dec 04 05:04:04 crc kubenswrapper[4806]: I1204 05:04:04.186362 4806 generic.go:334] "Generic (PLEG): container finished" podID="ed902041-2d66-4776-bd9e-5fc9f4f55048" containerID="75771ff539642ab63dc40c422e4057ba21b2dbf32411a7a122ca15d642bb2c07" exitCode=0 Dec 04 05:04:04 crc kubenswrapper[4806]: I1204 05:04:04.186880 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p55dr" event={"ID":"ed902041-2d66-4776-bd9e-5fc9f4f55048","Type":"ContainerDied","Data":"75771ff539642ab63dc40c422e4057ba21b2dbf32411a7a122ca15d642bb2c07"} Dec 04 05:04:05 crc kubenswrapper[4806]: I1204 05:04:05.199855 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p55dr" event={"ID":"ed902041-2d66-4776-bd9e-5fc9f4f55048","Type":"ContainerStarted","Data":"cea91e5e0c23b7f19f95489e3377cec38bfeed109361639c35c8d07f91100c14"} Dec 04 05:04:05 crc kubenswrapper[4806]: I1204 05:04:05.224764 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p55dr" podStartSLOduration=2.778194968 podStartE2EDuration="5.224747678s" podCreationTimestamp="2025-12-04 05:04:00 +0000 UTC" firstStartedPulling="2025-12-04 05:04:02.162404517 +0000 UTC m=+4157.020917445" lastFinishedPulling="2025-12-04 05:04:04.608957227 +0000 UTC m=+4159.467470155" observedRunningTime="2025-12-04 05:04:05.222009472 +0000 UTC m=+4160.080522410" watchObservedRunningTime="2025-12-04 05:04:05.224747678 +0000 UTC m=+4160.083260606" Dec 04 05:04:10 crc kubenswrapper[4806]: I1204 05:04:10.879753 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:10 crc kubenswrapper[4806]: I1204 05:04:10.880309 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:10 crc kubenswrapper[4806]: I1204 05:04:10.934198 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:11 crc kubenswrapper[4806]: I1204 05:04:11.321060 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:11 crc kubenswrapper[4806]: I1204 05:04:11.386766 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p55dr"] Dec 04 05:04:13 crc kubenswrapper[4806]: I1204 05:04:13.276565 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p55dr" podUID="ed902041-2d66-4776-bd9e-5fc9f4f55048" containerName="registry-server" containerID="cri-o://cea91e5e0c23b7f19f95489e3377cec38bfeed109361639c35c8d07f91100c14" gracePeriod=2 Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.032815 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.174306 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed902041-2d66-4776-bd9e-5fc9f4f55048-utilities\") pod \"ed902041-2d66-4776-bd9e-5fc9f4f55048\" (UID: \"ed902041-2d66-4776-bd9e-5fc9f4f55048\") " Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.174400 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25bsb\" (UniqueName: \"kubernetes.io/projected/ed902041-2d66-4776-bd9e-5fc9f4f55048-kube-api-access-25bsb\") pod \"ed902041-2d66-4776-bd9e-5fc9f4f55048\" (UID: \"ed902041-2d66-4776-bd9e-5fc9f4f55048\") " Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.174527 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed902041-2d66-4776-bd9e-5fc9f4f55048-catalog-content\") pod \"ed902041-2d66-4776-bd9e-5fc9f4f55048\" (UID: \"ed902041-2d66-4776-bd9e-5fc9f4f55048\") " Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.175689 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed902041-2d66-4776-bd9e-5fc9f4f55048-utilities" (OuterVolumeSpecName: "utilities") pod "ed902041-2d66-4776-bd9e-5fc9f4f55048" (UID: "ed902041-2d66-4776-bd9e-5fc9f4f55048"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.183852 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed902041-2d66-4776-bd9e-5fc9f4f55048-kube-api-access-25bsb" (OuterVolumeSpecName: "kube-api-access-25bsb") pod "ed902041-2d66-4776-bd9e-5fc9f4f55048" (UID: "ed902041-2d66-4776-bd9e-5fc9f4f55048"). InnerVolumeSpecName "kube-api-access-25bsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.235036 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed902041-2d66-4776-bd9e-5fc9f4f55048-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ed902041-2d66-4776-bd9e-5fc9f4f55048" (UID: "ed902041-2d66-4776-bd9e-5fc9f4f55048"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.279901 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed902041-2d66-4776-bd9e-5fc9f4f55048-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.279948 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed902041-2d66-4776-bd9e-5fc9f4f55048-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.279959 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25bsb\" (UniqueName: \"kubernetes.io/projected/ed902041-2d66-4776-bd9e-5fc9f4f55048-kube-api-access-25bsb\") on node \"crc\" DevicePath \"\"" Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.294885 4806 generic.go:334] "Generic (PLEG): container finished" podID="ed902041-2d66-4776-bd9e-5fc9f4f55048" containerID="cea91e5e0c23b7f19f95489e3377cec38bfeed109361639c35c8d07f91100c14" exitCode=0 Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.294949 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p55dr" event={"ID":"ed902041-2d66-4776-bd9e-5fc9f4f55048","Type":"ContainerDied","Data":"cea91e5e0c23b7f19f95489e3377cec38bfeed109361639c35c8d07f91100c14"} Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.294981 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p55dr" event={"ID":"ed902041-2d66-4776-bd9e-5fc9f4f55048","Type":"ContainerDied","Data":"702237eeefe301a21f27213c7f3f9dec607949eab301d8d67c6dad6eedfa3df4"} Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.295001 4806 scope.go:117] "RemoveContainer" containerID="cea91e5e0c23b7f19f95489e3377cec38bfeed109361639c35c8d07f91100c14" Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.295154 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p55dr" Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.343970 4806 scope.go:117] "RemoveContainer" containerID="75771ff539642ab63dc40c422e4057ba21b2dbf32411a7a122ca15d642bb2c07" Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.345047 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p55dr"] Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.357848 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p55dr"] Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.371819 4806 scope.go:117] "RemoveContainer" containerID="8df97461a0ad5ff6e070597601a66fca5f5c6078604701516ab6c7be9c9efdb6" Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.410171 4806 scope.go:117] "RemoveContainer" containerID="cea91e5e0c23b7f19f95489e3377cec38bfeed109361639c35c8d07f91100c14" Dec 04 05:04:14 crc kubenswrapper[4806]: E1204 05:04:14.410615 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cea91e5e0c23b7f19f95489e3377cec38bfeed109361639c35c8d07f91100c14\": container with ID starting with cea91e5e0c23b7f19f95489e3377cec38bfeed109361639c35c8d07f91100c14 not found: ID does not exist" containerID="cea91e5e0c23b7f19f95489e3377cec38bfeed109361639c35c8d07f91100c14" Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.410666 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cea91e5e0c23b7f19f95489e3377cec38bfeed109361639c35c8d07f91100c14"} err="failed to get container status \"cea91e5e0c23b7f19f95489e3377cec38bfeed109361639c35c8d07f91100c14\": rpc error: code = NotFound desc = could not find container \"cea91e5e0c23b7f19f95489e3377cec38bfeed109361639c35c8d07f91100c14\": container with ID starting with cea91e5e0c23b7f19f95489e3377cec38bfeed109361639c35c8d07f91100c14 not found: ID does not exist" Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.410701 4806 scope.go:117] "RemoveContainer" containerID="75771ff539642ab63dc40c422e4057ba21b2dbf32411a7a122ca15d642bb2c07" Dec 04 05:04:14 crc kubenswrapper[4806]: E1204 05:04:14.411049 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75771ff539642ab63dc40c422e4057ba21b2dbf32411a7a122ca15d642bb2c07\": container with ID starting with 75771ff539642ab63dc40c422e4057ba21b2dbf32411a7a122ca15d642bb2c07 not found: ID does not exist" containerID="75771ff539642ab63dc40c422e4057ba21b2dbf32411a7a122ca15d642bb2c07" Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.411081 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75771ff539642ab63dc40c422e4057ba21b2dbf32411a7a122ca15d642bb2c07"} err="failed to get container status \"75771ff539642ab63dc40c422e4057ba21b2dbf32411a7a122ca15d642bb2c07\": rpc error: code = NotFound desc = could not find container \"75771ff539642ab63dc40c422e4057ba21b2dbf32411a7a122ca15d642bb2c07\": container with ID starting with 75771ff539642ab63dc40c422e4057ba21b2dbf32411a7a122ca15d642bb2c07 not found: ID does not exist" Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.411097 4806 scope.go:117] "RemoveContainer" containerID="8df97461a0ad5ff6e070597601a66fca5f5c6078604701516ab6c7be9c9efdb6" Dec 04 05:04:14 crc kubenswrapper[4806]: E1204 05:04:14.411419 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8df97461a0ad5ff6e070597601a66fca5f5c6078604701516ab6c7be9c9efdb6\": container with ID starting with 8df97461a0ad5ff6e070597601a66fca5f5c6078604701516ab6c7be9c9efdb6 not found: ID does not exist" containerID="8df97461a0ad5ff6e070597601a66fca5f5c6078604701516ab6c7be9c9efdb6" Dec 04 05:04:14 crc kubenswrapper[4806]: I1204 05:04:14.411475 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8df97461a0ad5ff6e070597601a66fca5f5c6078604701516ab6c7be9c9efdb6"} err="failed to get container status \"8df97461a0ad5ff6e070597601a66fca5f5c6078604701516ab6c7be9c9efdb6\": rpc error: code = NotFound desc = could not find container \"8df97461a0ad5ff6e070597601a66fca5f5c6078604701516ab6c7be9c9efdb6\": container with ID starting with 8df97461a0ad5ff6e070597601a66fca5f5c6078604701516ab6c7be9c9efdb6 not found: ID does not exist" Dec 04 05:04:15 crc kubenswrapper[4806]: I1204 05:04:15.434307 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed902041-2d66-4776-bd9e-5fc9f4f55048" path="/var/lib/kubelet/pods/ed902041-2d66-4776-bd9e-5fc9f4f55048/volumes" Dec 04 05:04:27 crc kubenswrapper[4806]: I1204 05:04:27.047500 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:04:27 crc kubenswrapper[4806]: I1204 05:04:27.048115 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:04:57 crc kubenswrapper[4806]: I1204 05:04:57.046835 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:04:57 crc kubenswrapper[4806]: I1204 05:04:57.047462 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:05:27 crc kubenswrapper[4806]: I1204 05:05:27.047218 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:05:27 crc kubenswrapper[4806]: I1204 05:05:27.047847 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:05:27 crc kubenswrapper[4806]: I1204 05:05:27.047907 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 05:05:27 crc kubenswrapper[4806]: I1204 05:05:27.048795 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 05:05:27 crc kubenswrapper[4806]: I1204 05:05:27.048893 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" gracePeriod=600 Dec 04 05:05:27 crc kubenswrapper[4806]: E1204 05:05:27.170906 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:05:27 crc kubenswrapper[4806]: I1204 05:05:27.993566 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" exitCode=0 Dec 04 05:05:27 crc kubenswrapper[4806]: I1204 05:05:27.993644 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744"} Dec 04 05:05:27 crc kubenswrapper[4806]: I1204 05:05:27.993683 4806 scope.go:117] "RemoveContainer" containerID="b7a929db250bf8fd36191354c0529747a23a7c22591c6919b0c88de706949742" Dec 04 05:05:27 crc kubenswrapper[4806]: I1204 05:05:27.994748 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:05:27 crc kubenswrapper[4806]: E1204 05:05:27.995399 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:05:40 crc kubenswrapper[4806]: I1204 05:05:40.424326 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:05:40 crc kubenswrapper[4806]: E1204 05:05:40.425267 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.175770 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vc5b5"] Dec 04 05:05:48 crc kubenswrapper[4806]: E1204 05:05:48.176743 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed902041-2d66-4776-bd9e-5fc9f4f55048" containerName="registry-server" Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.176756 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed902041-2d66-4776-bd9e-5fc9f4f55048" containerName="registry-server" Dec 04 05:05:48 crc kubenswrapper[4806]: E1204 05:05:48.176792 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed902041-2d66-4776-bd9e-5fc9f4f55048" containerName="extract-content" Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.176798 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed902041-2d66-4776-bd9e-5fc9f4f55048" containerName="extract-content" Dec 04 05:05:48 crc kubenswrapper[4806]: E1204 05:05:48.176811 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed902041-2d66-4776-bd9e-5fc9f4f55048" containerName="extract-utilities" Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.176817 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed902041-2d66-4776-bd9e-5fc9f4f55048" containerName="extract-utilities" Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.177028 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed902041-2d66-4776-bd9e-5fc9f4f55048" containerName="registry-server" Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.178330 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.247685 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vc5b5"] Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.296020 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/341ff621-226f-4cf1-a31f-2464bfe65bd8-utilities\") pod \"redhat-marketplace-vc5b5\" (UID: \"341ff621-226f-4cf1-a31f-2464bfe65bd8\") " pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.296193 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzqdw\" (UniqueName: \"kubernetes.io/projected/341ff621-226f-4cf1-a31f-2464bfe65bd8-kube-api-access-pzqdw\") pod \"redhat-marketplace-vc5b5\" (UID: \"341ff621-226f-4cf1-a31f-2464bfe65bd8\") " pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.296276 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/341ff621-226f-4cf1-a31f-2464bfe65bd8-catalog-content\") pod \"redhat-marketplace-vc5b5\" (UID: \"341ff621-226f-4cf1-a31f-2464bfe65bd8\") " pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.397964 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzqdw\" (UniqueName: \"kubernetes.io/projected/341ff621-226f-4cf1-a31f-2464bfe65bd8-kube-api-access-pzqdw\") pod \"redhat-marketplace-vc5b5\" (UID: \"341ff621-226f-4cf1-a31f-2464bfe65bd8\") " pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.398052 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/341ff621-226f-4cf1-a31f-2464bfe65bd8-catalog-content\") pod \"redhat-marketplace-vc5b5\" (UID: \"341ff621-226f-4cf1-a31f-2464bfe65bd8\") " pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.398113 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/341ff621-226f-4cf1-a31f-2464bfe65bd8-utilities\") pod \"redhat-marketplace-vc5b5\" (UID: \"341ff621-226f-4cf1-a31f-2464bfe65bd8\") " pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.398690 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/341ff621-226f-4cf1-a31f-2464bfe65bd8-utilities\") pod \"redhat-marketplace-vc5b5\" (UID: \"341ff621-226f-4cf1-a31f-2464bfe65bd8\") " pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.398814 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/341ff621-226f-4cf1-a31f-2464bfe65bd8-catalog-content\") pod \"redhat-marketplace-vc5b5\" (UID: \"341ff621-226f-4cf1-a31f-2464bfe65bd8\") " pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.418046 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzqdw\" (UniqueName: \"kubernetes.io/projected/341ff621-226f-4cf1-a31f-2464bfe65bd8-kube-api-access-pzqdw\") pod \"redhat-marketplace-vc5b5\" (UID: \"341ff621-226f-4cf1-a31f-2464bfe65bd8\") " pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:05:48 crc kubenswrapper[4806]: I1204 05:05:48.510450 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:05:49 crc kubenswrapper[4806]: I1204 05:05:49.010595 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vc5b5"] Dec 04 05:05:49 crc kubenswrapper[4806]: I1204 05:05:49.191877 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vc5b5" event={"ID":"341ff621-226f-4cf1-a31f-2464bfe65bd8","Type":"ContainerStarted","Data":"aa19cbd0915a26fee89a49eb9e7d14dc5d2166907ac9e97e819c75ddd1dbcca2"} Dec 04 05:05:50 crc kubenswrapper[4806]: I1204 05:05:50.201390 4806 generic.go:334] "Generic (PLEG): container finished" podID="341ff621-226f-4cf1-a31f-2464bfe65bd8" containerID="1cb10db5b1f173c0c6bac5d74421e7048377b37bf85e2e6e5bb12c661778aadf" exitCode=0 Dec 04 05:05:50 crc kubenswrapper[4806]: I1204 05:05:50.201452 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vc5b5" event={"ID":"341ff621-226f-4cf1-a31f-2464bfe65bd8","Type":"ContainerDied","Data":"1cb10db5b1f173c0c6bac5d74421e7048377b37bf85e2e6e5bb12c661778aadf"} Dec 04 05:05:50 crc kubenswrapper[4806]: I1204 05:05:50.204236 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 05:05:51 crc kubenswrapper[4806]: I1204 05:05:51.212027 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vc5b5" event={"ID":"341ff621-226f-4cf1-a31f-2464bfe65bd8","Type":"ContainerStarted","Data":"056189c22a9ad3fabb12a6eb3cb51a9169262f008fb4d43cb7b588a10e72db27"} Dec 04 05:05:52 crc kubenswrapper[4806]: I1204 05:05:52.222904 4806 generic.go:334] "Generic (PLEG): container finished" podID="341ff621-226f-4cf1-a31f-2464bfe65bd8" containerID="056189c22a9ad3fabb12a6eb3cb51a9169262f008fb4d43cb7b588a10e72db27" exitCode=0 Dec 04 05:05:52 crc kubenswrapper[4806]: I1204 05:05:52.223015 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vc5b5" event={"ID":"341ff621-226f-4cf1-a31f-2464bfe65bd8","Type":"ContainerDied","Data":"056189c22a9ad3fabb12a6eb3cb51a9169262f008fb4d43cb7b588a10e72db27"} Dec 04 05:05:53 crc kubenswrapper[4806]: I1204 05:05:53.425227 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:05:53 crc kubenswrapper[4806]: E1204 05:05:53.425851 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:05:55 crc kubenswrapper[4806]: I1204 05:05:55.251956 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vc5b5" event={"ID":"341ff621-226f-4cf1-a31f-2464bfe65bd8","Type":"ContainerStarted","Data":"c6bb2aa00218988e470a129ffa438f3fe4e7783fadc4b10533afce5160929a5e"} Dec 04 05:05:55 crc kubenswrapper[4806]: I1204 05:05:55.276815 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vc5b5" podStartSLOduration=2.685610662 podStartE2EDuration="7.276793341s" podCreationTimestamp="2025-12-04 05:05:48 +0000 UTC" firstStartedPulling="2025-12-04 05:05:50.20395589 +0000 UTC m=+4265.062468818" lastFinishedPulling="2025-12-04 05:05:54.795138569 +0000 UTC m=+4269.653651497" observedRunningTime="2025-12-04 05:05:55.270771051 +0000 UTC m=+4270.129283969" watchObservedRunningTime="2025-12-04 05:05:55.276793341 +0000 UTC m=+4270.135306269" Dec 04 05:05:58 crc kubenswrapper[4806]: I1204 05:05:58.511421 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:05:58 crc kubenswrapper[4806]: I1204 05:05:58.512381 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:05:58 crc kubenswrapper[4806]: I1204 05:05:58.557247 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:06:05 crc kubenswrapper[4806]: I1204 05:06:05.429589 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:06:05 crc kubenswrapper[4806]: E1204 05:06:05.430262 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:06:08 crc kubenswrapper[4806]: I1204 05:06:08.561771 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:06:08 crc kubenswrapper[4806]: I1204 05:06:08.617118 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vc5b5"] Dec 04 05:06:09 crc kubenswrapper[4806]: I1204 05:06:09.383511 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vc5b5" podUID="341ff621-226f-4cf1-a31f-2464bfe65bd8" containerName="registry-server" containerID="cri-o://c6bb2aa00218988e470a129ffa438f3fe4e7783fadc4b10533afce5160929a5e" gracePeriod=2 Dec 04 05:06:09 crc kubenswrapper[4806]: I1204 05:06:09.974668 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.018547 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/341ff621-226f-4cf1-a31f-2464bfe65bd8-utilities\") pod \"341ff621-226f-4cf1-a31f-2464bfe65bd8\" (UID: \"341ff621-226f-4cf1-a31f-2464bfe65bd8\") " Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.018957 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzqdw\" (UniqueName: \"kubernetes.io/projected/341ff621-226f-4cf1-a31f-2464bfe65bd8-kube-api-access-pzqdw\") pod \"341ff621-226f-4cf1-a31f-2464bfe65bd8\" (UID: \"341ff621-226f-4cf1-a31f-2464bfe65bd8\") " Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.019129 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/341ff621-226f-4cf1-a31f-2464bfe65bd8-catalog-content\") pod \"341ff621-226f-4cf1-a31f-2464bfe65bd8\" (UID: \"341ff621-226f-4cf1-a31f-2464bfe65bd8\") " Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.019866 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/341ff621-226f-4cf1-a31f-2464bfe65bd8-utilities" (OuterVolumeSpecName: "utilities") pod "341ff621-226f-4cf1-a31f-2464bfe65bd8" (UID: "341ff621-226f-4cf1-a31f-2464bfe65bd8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.042910 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/341ff621-226f-4cf1-a31f-2464bfe65bd8-kube-api-access-pzqdw" (OuterVolumeSpecName: "kube-api-access-pzqdw") pod "341ff621-226f-4cf1-a31f-2464bfe65bd8" (UID: "341ff621-226f-4cf1-a31f-2464bfe65bd8"). InnerVolumeSpecName "kube-api-access-pzqdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.045311 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/341ff621-226f-4cf1-a31f-2464bfe65bd8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "341ff621-226f-4cf1-a31f-2464bfe65bd8" (UID: "341ff621-226f-4cf1-a31f-2464bfe65bd8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.122261 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzqdw\" (UniqueName: \"kubernetes.io/projected/341ff621-226f-4cf1-a31f-2464bfe65bd8-kube-api-access-pzqdw\") on node \"crc\" DevicePath \"\"" Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.122322 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/341ff621-226f-4cf1-a31f-2464bfe65bd8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.122334 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/341ff621-226f-4cf1-a31f-2464bfe65bd8-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.393824 4806 generic.go:334] "Generic (PLEG): container finished" podID="341ff621-226f-4cf1-a31f-2464bfe65bd8" containerID="c6bb2aa00218988e470a129ffa438f3fe4e7783fadc4b10533afce5160929a5e" exitCode=0 Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.393885 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vc5b5" Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.393913 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vc5b5" event={"ID":"341ff621-226f-4cf1-a31f-2464bfe65bd8","Type":"ContainerDied","Data":"c6bb2aa00218988e470a129ffa438f3fe4e7783fadc4b10533afce5160929a5e"} Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.394223 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vc5b5" event={"ID":"341ff621-226f-4cf1-a31f-2464bfe65bd8","Type":"ContainerDied","Data":"aa19cbd0915a26fee89a49eb9e7d14dc5d2166907ac9e97e819c75ddd1dbcca2"} Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.394254 4806 scope.go:117] "RemoveContainer" containerID="c6bb2aa00218988e470a129ffa438f3fe4e7783fadc4b10533afce5160929a5e" Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.426314 4806 scope.go:117] "RemoveContainer" containerID="056189c22a9ad3fabb12a6eb3cb51a9169262f008fb4d43cb7b588a10e72db27" Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.433036 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vc5b5"] Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.450114 4806 scope.go:117] "RemoveContainer" containerID="1cb10db5b1f173c0c6bac5d74421e7048377b37bf85e2e6e5bb12c661778aadf" Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.471140 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vc5b5"] Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.496516 4806 scope.go:117] "RemoveContainer" containerID="c6bb2aa00218988e470a129ffa438f3fe4e7783fadc4b10533afce5160929a5e" Dec 04 05:06:10 crc kubenswrapper[4806]: E1204 05:06:10.499482 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6bb2aa00218988e470a129ffa438f3fe4e7783fadc4b10533afce5160929a5e\": container with ID starting with c6bb2aa00218988e470a129ffa438f3fe4e7783fadc4b10533afce5160929a5e not found: ID does not exist" containerID="c6bb2aa00218988e470a129ffa438f3fe4e7783fadc4b10533afce5160929a5e" Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.499639 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6bb2aa00218988e470a129ffa438f3fe4e7783fadc4b10533afce5160929a5e"} err="failed to get container status \"c6bb2aa00218988e470a129ffa438f3fe4e7783fadc4b10533afce5160929a5e\": rpc error: code = NotFound desc = could not find container \"c6bb2aa00218988e470a129ffa438f3fe4e7783fadc4b10533afce5160929a5e\": container with ID starting with c6bb2aa00218988e470a129ffa438f3fe4e7783fadc4b10533afce5160929a5e not found: ID does not exist" Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.499719 4806 scope.go:117] "RemoveContainer" containerID="056189c22a9ad3fabb12a6eb3cb51a9169262f008fb4d43cb7b588a10e72db27" Dec 04 05:06:10 crc kubenswrapper[4806]: E1204 05:06:10.500694 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"056189c22a9ad3fabb12a6eb3cb51a9169262f008fb4d43cb7b588a10e72db27\": container with ID starting with 056189c22a9ad3fabb12a6eb3cb51a9169262f008fb4d43cb7b588a10e72db27 not found: ID does not exist" containerID="056189c22a9ad3fabb12a6eb3cb51a9169262f008fb4d43cb7b588a10e72db27" Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.500786 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"056189c22a9ad3fabb12a6eb3cb51a9169262f008fb4d43cb7b588a10e72db27"} err="failed to get container status \"056189c22a9ad3fabb12a6eb3cb51a9169262f008fb4d43cb7b588a10e72db27\": rpc error: code = NotFound desc = could not find container \"056189c22a9ad3fabb12a6eb3cb51a9169262f008fb4d43cb7b588a10e72db27\": container with ID starting with 056189c22a9ad3fabb12a6eb3cb51a9169262f008fb4d43cb7b588a10e72db27 not found: ID does not exist" Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.500863 4806 scope.go:117] "RemoveContainer" containerID="1cb10db5b1f173c0c6bac5d74421e7048377b37bf85e2e6e5bb12c661778aadf" Dec 04 05:06:10 crc kubenswrapper[4806]: E1204 05:06:10.501912 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cb10db5b1f173c0c6bac5d74421e7048377b37bf85e2e6e5bb12c661778aadf\": container with ID starting with 1cb10db5b1f173c0c6bac5d74421e7048377b37bf85e2e6e5bb12c661778aadf not found: ID does not exist" containerID="1cb10db5b1f173c0c6bac5d74421e7048377b37bf85e2e6e5bb12c661778aadf" Dec 04 05:06:10 crc kubenswrapper[4806]: I1204 05:06:10.502010 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cb10db5b1f173c0c6bac5d74421e7048377b37bf85e2e6e5bb12c661778aadf"} err="failed to get container status \"1cb10db5b1f173c0c6bac5d74421e7048377b37bf85e2e6e5bb12c661778aadf\": rpc error: code = NotFound desc = could not find container \"1cb10db5b1f173c0c6bac5d74421e7048377b37bf85e2e6e5bb12c661778aadf\": container with ID starting with 1cb10db5b1f173c0c6bac5d74421e7048377b37bf85e2e6e5bb12c661778aadf not found: ID does not exist" Dec 04 05:06:11 crc kubenswrapper[4806]: I1204 05:06:11.434818 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="341ff621-226f-4cf1-a31f-2464bfe65bd8" path="/var/lib/kubelet/pods/341ff621-226f-4cf1-a31f-2464bfe65bd8/volumes" Dec 04 05:06:17 crc kubenswrapper[4806]: I1204 05:06:17.424518 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:06:17 crc kubenswrapper[4806]: E1204 05:06:17.425381 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:06:31 crc kubenswrapper[4806]: I1204 05:06:31.423370 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:06:31 crc kubenswrapper[4806]: E1204 05:06:31.424140 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:06:43 crc kubenswrapper[4806]: I1204 05:06:43.424493 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:06:43 crc kubenswrapper[4806]: E1204 05:06:43.425337 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:06:55 crc kubenswrapper[4806]: I1204 05:06:55.430510 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:06:55 crc kubenswrapper[4806]: E1204 05:06:55.431411 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:07:06 crc kubenswrapper[4806]: I1204 05:07:06.423896 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:07:06 crc kubenswrapper[4806]: E1204 05:07:06.424703 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:07:17 crc kubenswrapper[4806]: I1204 05:07:17.423968 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:07:17 crc kubenswrapper[4806]: E1204 05:07:17.424786 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:07:32 crc kubenswrapper[4806]: I1204 05:07:32.423968 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:07:32 crc kubenswrapper[4806]: E1204 05:07:32.425228 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:07:44 crc kubenswrapper[4806]: I1204 05:07:44.424122 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:07:44 crc kubenswrapper[4806]: E1204 05:07:44.424978 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:07:57 crc kubenswrapper[4806]: I1204 05:07:57.424010 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:07:57 crc kubenswrapper[4806]: E1204 05:07:57.424726 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:08:09 crc kubenswrapper[4806]: I1204 05:08:09.424283 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:08:09 crc kubenswrapper[4806]: E1204 05:08:09.425064 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:08:20 crc kubenswrapper[4806]: I1204 05:08:20.424123 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:08:20 crc kubenswrapper[4806]: E1204 05:08:20.425164 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:08:34 crc kubenswrapper[4806]: I1204 05:08:34.424134 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:08:34 crc kubenswrapper[4806]: E1204 05:08:34.424907 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:08:49 crc kubenswrapper[4806]: I1204 05:08:49.425324 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:08:49 crc kubenswrapper[4806]: E1204 05:08:49.426139 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:09:01 crc kubenswrapper[4806]: I1204 05:09:01.424692 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:09:01 crc kubenswrapper[4806]: E1204 05:09:01.426741 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:09:12 crc kubenswrapper[4806]: I1204 05:09:12.423843 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:09:12 crc kubenswrapper[4806]: E1204 05:09:12.424576 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:09:23 crc kubenswrapper[4806]: I1204 05:09:23.425320 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:09:23 crc kubenswrapper[4806]: E1204 05:09:23.426950 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:09:37 crc kubenswrapper[4806]: I1204 05:09:37.424318 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:09:37 crc kubenswrapper[4806]: E1204 05:09:37.425141 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:09:40 crc kubenswrapper[4806]: I1204 05:09:40.769130 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k97d6"] Dec 04 05:09:40 crc kubenswrapper[4806]: E1204 05:09:40.776678 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="341ff621-226f-4cf1-a31f-2464bfe65bd8" containerName="extract-content" Dec 04 05:09:40 crc kubenswrapper[4806]: I1204 05:09:40.776705 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="341ff621-226f-4cf1-a31f-2464bfe65bd8" containerName="extract-content" Dec 04 05:09:40 crc kubenswrapper[4806]: E1204 05:09:40.776755 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="341ff621-226f-4cf1-a31f-2464bfe65bd8" containerName="extract-utilities" Dec 04 05:09:40 crc kubenswrapper[4806]: I1204 05:09:40.776765 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="341ff621-226f-4cf1-a31f-2464bfe65bd8" containerName="extract-utilities" Dec 04 05:09:40 crc kubenswrapper[4806]: E1204 05:09:40.776776 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="341ff621-226f-4cf1-a31f-2464bfe65bd8" containerName="registry-server" Dec 04 05:09:40 crc kubenswrapper[4806]: I1204 05:09:40.776782 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="341ff621-226f-4cf1-a31f-2464bfe65bd8" containerName="registry-server" Dec 04 05:09:40 crc kubenswrapper[4806]: I1204 05:09:40.777024 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="341ff621-226f-4cf1-a31f-2464bfe65bd8" containerName="registry-server" Dec 04 05:09:40 crc kubenswrapper[4806]: I1204 05:09:40.778736 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:40 crc kubenswrapper[4806]: I1204 05:09:40.794950 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k97d6"] Dec 04 05:09:40 crc kubenswrapper[4806]: I1204 05:09:40.871412 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a583375-a2ee-469c-a6c3-0c3315fc1833-catalog-content\") pod \"certified-operators-k97d6\" (UID: \"7a583375-a2ee-469c-a6c3-0c3315fc1833\") " pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:40 crc kubenswrapper[4806]: I1204 05:09:40.871481 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a583375-a2ee-469c-a6c3-0c3315fc1833-utilities\") pod \"certified-operators-k97d6\" (UID: \"7a583375-a2ee-469c-a6c3-0c3315fc1833\") " pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:40 crc kubenswrapper[4806]: I1204 05:09:40.871592 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf4r5\" (UniqueName: \"kubernetes.io/projected/7a583375-a2ee-469c-a6c3-0c3315fc1833-kube-api-access-tf4r5\") pod \"certified-operators-k97d6\" (UID: \"7a583375-a2ee-469c-a6c3-0c3315fc1833\") " pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:40 crc kubenswrapper[4806]: I1204 05:09:40.973683 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf4r5\" (UniqueName: \"kubernetes.io/projected/7a583375-a2ee-469c-a6c3-0c3315fc1833-kube-api-access-tf4r5\") pod \"certified-operators-k97d6\" (UID: \"7a583375-a2ee-469c-a6c3-0c3315fc1833\") " pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:40 crc kubenswrapper[4806]: I1204 05:09:40.974038 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a583375-a2ee-469c-a6c3-0c3315fc1833-catalog-content\") pod \"certified-operators-k97d6\" (UID: \"7a583375-a2ee-469c-a6c3-0c3315fc1833\") " pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:40 crc kubenswrapper[4806]: I1204 05:09:40.974128 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a583375-a2ee-469c-a6c3-0c3315fc1833-utilities\") pod \"certified-operators-k97d6\" (UID: \"7a583375-a2ee-469c-a6c3-0c3315fc1833\") " pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:40 crc kubenswrapper[4806]: I1204 05:09:40.974778 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a583375-a2ee-469c-a6c3-0c3315fc1833-catalog-content\") pod \"certified-operators-k97d6\" (UID: \"7a583375-a2ee-469c-a6c3-0c3315fc1833\") " pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:40 crc kubenswrapper[4806]: I1204 05:09:40.974789 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a583375-a2ee-469c-a6c3-0c3315fc1833-utilities\") pod \"certified-operators-k97d6\" (UID: \"7a583375-a2ee-469c-a6c3-0c3315fc1833\") " pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:41 crc kubenswrapper[4806]: I1204 05:09:41.000077 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf4r5\" (UniqueName: \"kubernetes.io/projected/7a583375-a2ee-469c-a6c3-0c3315fc1833-kube-api-access-tf4r5\") pod \"certified-operators-k97d6\" (UID: \"7a583375-a2ee-469c-a6c3-0c3315fc1833\") " pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:41 crc kubenswrapper[4806]: I1204 05:09:41.100996 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:41 crc kubenswrapper[4806]: I1204 05:09:41.872137 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k97d6"] Dec 04 05:09:41 crc kubenswrapper[4806]: W1204 05:09:41.883889 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a583375_a2ee_469c_a6c3_0c3315fc1833.slice/crio-3f8c32faac0fda15be7dcb1aed19787eba0015e7e84518ef2ff574060bcb6eb5 WatchSource:0}: Error finding container 3f8c32faac0fda15be7dcb1aed19787eba0015e7e84518ef2ff574060bcb6eb5: Status 404 returned error can't find the container with id 3f8c32faac0fda15be7dcb1aed19787eba0015e7e84518ef2ff574060bcb6eb5 Dec 04 05:09:42 crc kubenswrapper[4806]: I1204 05:09:42.290448 4806 generic.go:334] "Generic (PLEG): container finished" podID="7a583375-a2ee-469c-a6c3-0c3315fc1833" containerID="9c546f72c39d7cb2b49675c8270e7763d844042ce140db3323bb3cdc1966b53e" exitCode=0 Dec 04 05:09:42 crc kubenswrapper[4806]: I1204 05:09:42.291045 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k97d6" event={"ID":"7a583375-a2ee-469c-a6c3-0c3315fc1833","Type":"ContainerDied","Data":"9c546f72c39d7cb2b49675c8270e7763d844042ce140db3323bb3cdc1966b53e"} Dec 04 05:09:42 crc kubenswrapper[4806]: I1204 05:09:42.291160 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k97d6" event={"ID":"7a583375-a2ee-469c-a6c3-0c3315fc1833","Type":"ContainerStarted","Data":"3f8c32faac0fda15be7dcb1aed19787eba0015e7e84518ef2ff574060bcb6eb5"} Dec 04 05:09:44 crc kubenswrapper[4806]: I1204 05:09:44.310770 4806 generic.go:334] "Generic (PLEG): container finished" podID="7a583375-a2ee-469c-a6c3-0c3315fc1833" containerID="5bcc9891f3819a1694d2d12220b219eb3b972be4c6771c93161f8ec15a4b2640" exitCode=0 Dec 04 05:09:44 crc kubenswrapper[4806]: I1204 05:09:44.310812 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k97d6" event={"ID":"7a583375-a2ee-469c-a6c3-0c3315fc1833","Type":"ContainerDied","Data":"5bcc9891f3819a1694d2d12220b219eb3b972be4c6771c93161f8ec15a4b2640"} Dec 04 05:09:45 crc kubenswrapper[4806]: I1204 05:09:45.321852 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k97d6" event={"ID":"7a583375-a2ee-469c-a6c3-0c3315fc1833","Type":"ContainerStarted","Data":"68ec4961b7f01f413e87df52e48f6e263479c5774232878f57f0e81d257a3800"} Dec 04 05:09:45 crc kubenswrapper[4806]: I1204 05:09:45.352188 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k97d6" podStartSLOduration=2.905768639 podStartE2EDuration="5.352164851s" podCreationTimestamp="2025-12-04 05:09:40 +0000 UTC" firstStartedPulling="2025-12-04 05:09:42.293557818 +0000 UTC m=+4497.152070746" lastFinishedPulling="2025-12-04 05:09:44.73995403 +0000 UTC m=+4499.598466958" observedRunningTime="2025-12-04 05:09:45.346915736 +0000 UTC m=+4500.205428654" watchObservedRunningTime="2025-12-04 05:09:45.352164851 +0000 UTC m=+4500.210677779" Dec 04 05:09:51 crc kubenswrapper[4806]: I1204 05:09:51.101883 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:51 crc kubenswrapper[4806]: I1204 05:09:51.102456 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:51 crc kubenswrapper[4806]: I1204 05:09:51.151144 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:51 crc kubenswrapper[4806]: I1204 05:09:51.424069 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:09:51 crc kubenswrapper[4806]: E1204 05:09:51.424369 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:09:51 crc kubenswrapper[4806]: I1204 05:09:51.435774 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:51 crc kubenswrapper[4806]: I1204 05:09:51.498713 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k97d6"] Dec 04 05:09:53 crc kubenswrapper[4806]: I1204 05:09:53.392755 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k97d6" podUID="7a583375-a2ee-469c-a6c3-0c3315fc1833" containerName="registry-server" containerID="cri-o://68ec4961b7f01f413e87df52e48f6e263479c5774232878f57f0e81d257a3800" gracePeriod=2 Dec 04 05:09:53 crc kubenswrapper[4806]: I1204 05:09:53.894853 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:53 crc kubenswrapper[4806]: I1204 05:09:53.936467 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a583375-a2ee-469c-a6c3-0c3315fc1833-utilities\") pod \"7a583375-a2ee-469c-a6c3-0c3315fc1833\" (UID: \"7a583375-a2ee-469c-a6c3-0c3315fc1833\") " Dec 04 05:09:53 crc kubenswrapper[4806]: I1204 05:09:53.936507 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf4r5\" (UniqueName: \"kubernetes.io/projected/7a583375-a2ee-469c-a6c3-0c3315fc1833-kube-api-access-tf4r5\") pod \"7a583375-a2ee-469c-a6c3-0c3315fc1833\" (UID: \"7a583375-a2ee-469c-a6c3-0c3315fc1833\") " Dec 04 05:09:53 crc kubenswrapper[4806]: I1204 05:09:53.936567 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a583375-a2ee-469c-a6c3-0c3315fc1833-catalog-content\") pod \"7a583375-a2ee-469c-a6c3-0c3315fc1833\" (UID: \"7a583375-a2ee-469c-a6c3-0c3315fc1833\") " Dec 04 05:09:53 crc kubenswrapper[4806]: I1204 05:09:53.938806 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a583375-a2ee-469c-a6c3-0c3315fc1833-utilities" (OuterVolumeSpecName: "utilities") pod "7a583375-a2ee-469c-a6c3-0c3315fc1833" (UID: "7a583375-a2ee-469c-a6c3-0c3315fc1833"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:09:53 crc kubenswrapper[4806]: I1204 05:09:53.962747 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a583375-a2ee-469c-a6c3-0c3315fc1833-kube-api-access-tf4r5" (OuterVolumeSpecName: "kube-api-access-tf4r5") pod "7a583375-a2ee-469c-a6c3-0c3315fc1833" (UID: "7a583375-a2ee-469c-a6c3-0c3315fc1833"). InnerVolumeSpecName "kube-api-access-tf4r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.003065 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a583375-a2ee-469c-a6c3-0c3315fc1833-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a583375-a2ee-469c-a6c3-0c3315fc1833" (UID: "7a583375-a2ee-469c-a6c3-0c3315fc1833"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.039360 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a583375-a2ee-469c-a6c3-0c3315fc1833-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.039582 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf4r5\" (UniqueName: \"kubernetes.io/projected/7a583375-a2ee-469c-a6c3-0c3315fc1833-kube-api-access-tf4r5\") on node \"crc\" DevicePath \"\"" Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.039636 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a583375-a2ee-469c-a6c3-0c3315fc1833-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.402594 4806 generic.go:334] "Generic (PLEG): container finished" podID="7a583375-a2ee-469c-a6c3-0c3315fc1833" containerID="68ec4961b7f01f413e87df52e48f6e263479c5774232878f57f0e81d257a3800" exitCode=0 Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.402639 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k97d6" Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.402641 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k97d6" event={"ID":"7a583375-a2ee-469c-a6c3-0c3315fc1833","Type":"ContainerDied","Data":"68ec4961b7f01f413e87df52e48f6e263479c5774232878f57f0e81d257a3800"} Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.402766 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k97d6" event={"ID":"7a583375-a2ee-469c-a6c3-0c3315fc1833","Type":"ContainerDied","Data":"3f8c32faac0fda15be7dcb1aed19787eba0015e7e84518ef2ff574060bcb6eb5"} Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.402796 4806 scope.go:117] "RemoveContainer" containerID="68ec4961b7f01f413e87df52e48f6e263479c5774232878f57f0e81d257a3800" Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.421338 4806 scope.go:117] "RemoveContainer" containerID="5bcc9891f3819a1694d2d12220b219eb3b972be4c6771c93161f8ec15a4b2640" Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.438024 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k97d6"] Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.447445 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k97d6"] Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.464345 4806 scope.go:117] "RemoveContainer" containerID="9c546f72c39d7cb2b49675c8270e7763d844042ce140db3323bb3cdc1966b53e" Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.501188 4806 scope.go:117] "RemoveContainer" containerID="68ec4961b7f01f413e87df52e48f6e263479c5774232878f57f0e81d257a3800" Dec 04 05:09:54 crc kubenswrapper[4806]: E1204 05:09:54.501693 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68ec4961b7f01f413e87df52e48f6e263479c5774232878f57f0e81d257a3800\": container with ID starting with 68ec4961b7f01f413e87df52e48f6e263479c5774232878f57f0e81d257a3800 not found: ID does not exist" containerID="68ec4961b7f01f413e87df52e48f6e263479c5774232878f57f0e81d257a3800" Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.501748 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68ec4961b7f01f413e87df52e48f6e263479c5774232878f57f0e81d257a3800"} err="failed to get container status \"68ec4961b7f01f413e87df52e48f6e263479c5774232878f57f0e81d257a3800\": rpc error: code = NotFound desc = could not find container \"68ec4961b7f01f413e87df52e48f6e263479c5774232878f57f0e81d257a3800\": container with ID starting with 68ec4961b7f01f413e87df52e48f6e263479c5774232878f57f0e81d257a3800 not found: ID does not exist" Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.501775 4806 scope.go:117] "RemoveContainer" containerID="5bcc9891f3819a1694d2d12220b219eb3b972be4c6771c93161f8ec15a4b2640" Dec 04 05:09:54 crc kubenswrapper[4806]: E1204 05:09:54.502069 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bcc9891f3819a1694d2d12220b219eb3b972be4c6771c93161f8ec15a4b2640\": container with ID starting with 5bcc9891f3819a1694d2d12220b219eb3b972be4c6771c93161f8ec15a4b2640 not found: ID does not exist" containerID="5bcc9891f3819a1694d2d12220b219eb3b972be4c6771c93161f8ec15a4b2640" Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.502454 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bcc9891f3819a1694d2d12220b219eb3b972be4c6771c93161f8ec15a4b2640"} err="failed to get container status \"5bcc9891f3819a1694d2d12220b219eb3b972be4c6771c93161f8ec15a4b2640\": rpc error: code = NotFound desc = could not find container \"5bcc9891f3819a1694d2d12220b219eb3b972be4c6771c93161f8ec15a4b2640\": container with ID starting with 5bcc9891f3819a1694d2d12220b219eb3b972be4c6771c93161f8ec15a4b2640 not found: ID does not exist" Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.502481 4806 scope.go:117] "RemoveContainer" containerID="9c546f72c39d7cb2b49675c8270e7763d844042ce140db3323bb3cdc1966b53e" Dec 04 05:09:54 crc kubenswrapper[4806]: E1204 05:09:54.502834 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c546f72c39d7cb2b49675c8270e7763d844042ce140db3323bb3cdc1966b53e\": container with ID starting with 9c546f72c39d7cb2b49675c8270e7763d844042ce140db3323bb3cdc1966b53e not found: ID does not exist" containerID="9c546f72c39d7cb2b49675c8270e7763d844042ce140db3323bb3cdc1966b53e" Dec 04 05:09:54 crc kubenswrapper[4806]: I1204 05:09:54.502877 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c546f72c39d7cb2b49675c8270e7763d844042ce140db3323bb3cdc1966b53e"} err="failed to get container status \"9c546f72c39d7cb2b49675c8270e7763d844042ce140db3323bb3cdc1966b53e\": rpc error: code = NotFound desc = could not find container \"9c546f72c39d7cb2b49675c8270e7763d844042ce140db3323bb3cdc1966b53e\": container with ID starting with 9c546f72c39d7cb2b49675c8270e7763d844042ce140db3323bb3cdc1966b53e not found: ID does not exist" Dec 04 05:09:55 crc kubenswrapper[4806]: I1204 05:09:55.438592 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a583375-a2ee-469c-a6c3-0c3315fc1833" path="/var/lib/kubelet/pods/7a583375-a2ee-469c-a6c3-0c3315fc1833/volumes" Dec 04 05:10:03 crc kubenswrapper[4806]: I1204 05:10:03.423397 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:10:03 crc kubenswrapper[4806]: E1204 05:10:03.424118 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:10:17 crc kubenswrapper[4806]: I1204 05:10:17.423892 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:10:17 crc kubenswrapper[4806]: E1204 05:10:17.425183 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:10:29 crc kubenswrapper[4806]: I1204 05:10:29.423908 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:10:30 crc kubenswrapper[4806]: I1204 05:10:30.724337 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"a1e6c4d9918f0691a954f6d7dd837700fe9fc560ddcdf3bc751c5b70d3e912e2"} Dec 04 05:12:36 crc kubenswrapper[4806]: I1204 05:12:36.822220 4806 generic.go:334] "Generic (PLEG): container finished" podID="319a7357-70c6-4b54-8e97-9860dd0618e7" containerID="53bbe8fc4a8361c9961c95aec4994de66b7c8f7d18efb9bbe065be6e3ba94996" exitCode=0 Dec 04 05:12:36 crc kubenswrapper[4806]: I1204 05:12:36.822338 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"319a7357-70c6-4b54-8e97-9860dd0618e7","Type":"ContainerDied","Data":"53bbe8fc4a8361c9961c95aec4994de66b7c8f7d18efb9bbe065be6e3ba94996"} Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.326625 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.424187 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"319a7357-70c6-4b54-8e97-9860dd0618e7\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.424548 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-ca-certs\") pod \"319a7357-70c6-4b54-8e97-9860dd0618e7\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.424611 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/319a7357-70c6-4b54-8e97-9860dd0618e7-test-operator-ephemeral-temporary\") pod \"319a7357-70c6-4b54-8e97-9860dd0618e7\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.424673 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/319a7357-70c6-4b54-8e97-9860dd0618e7-test-operator-ephemeral-workdir\") pod \"319a7357-70c6-4b54-8e97-9860dd0618e7\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.424719 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njg9d\" (UniqueName: \"kubernetes.io/projected/319a7357-70c6-4b54-8e97-9860dd0618e7-kube-api-access-njg9d\") pod \"319a7357-70c6-4b54-8e97-9860dd0618e7\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.424776 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-ssh-key\") pod \"319a7357-70c6-4b54-8e97-9860dd0618e7\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.424811 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/319a7357-70c6-4b54-8e97-9860dd0618e7-openstack-config\") pod \"319a7357-70c6-4b54-8e97-9860dd0618e7\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.424864 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-openstack-config-secret\") pod \"319a7357-70c6-4b54-8e97-9860dd0618e7\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.424913 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/319a7357-70c6-4b54-8e97-9860dd0618e7-config-data\") pod \"319a7357-70c6-4b54-8e97-9860dd0618e7\" (UID: \"319a7357-70c6-4b54-8e97-9860dd0618e7\") " Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.426744 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/319a7357-70c6-4b54-8e97-9860dd0618e7-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "319a7357-70c6-4b54-8e97-9860dd0618e7" (UID: "319a7357-70c6-4b54-8e97-9860dd0618e7"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.427470 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/319a7357-70c6-4b54-8e97-9860dd0618e7-config-data" (OuterVolumeSpecName: "config-data") pod "319a7357-70c6-4b54-8e97-9860dd0618e7" (UID: "319a7357-70c6-4b54-8e97-9860dd0618e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.431139 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/319a7357-70c6-4b54-8e97-9860dd0618e7-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "319a7357-70c6-4b54-8e97-9860dd0618e7" (UID: "319a7357-70c6-4b54-8e97-9860dd0618e7"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.431757 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "test-operator-logs") pod "319a7357-70c6-4b54-8e97-9860dd0618e7" (UID: "319a7357-70c6-4b54-8e97-9860dd0618e7"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.445609 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/319a7357-70c6-4b54-8e97-9860dd0618e7-kube-api-access-njg9d" (OuterVolumeSpecName: "kube-api-access-njg9d") pod "319a7357-70c6-4b54-8e97-9860dd0618e7" (UID: "319a7357-70c6-4b54-8e97-9860dd0618e7"). InnerVolumeSpecName "kube-api-access-njg9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.460196 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "319a7357-70c6-4b54-8e97-9860dd0618e7" (UID: "319a7357-70c6-4b54-8e97-9860dd0618e7"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.462293 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "319a7357-70c6-4b54-8e97-9860dd0618e7" (UID: "319a7357-70c6-4b54-8e97-9860dd0618e7"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.469729 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "319a7357-70c6-4b54-8e97-9860dd0618e7" (UID: "319a7357-70c6-4b54-8e97-9860dd0618e7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.499949 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/319a7357-70c6-4b54-8e97-9860dd0618e7-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "319a7357-70c6-4b54-8e97-9860dd0618e7" (UID: "319a7357-70c6-4b54-8e97-9860dd0618e7"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.527400 4806 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/319a7357-70c6-4b54-8e97-9860dd0618e7-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.527536 4806 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/319a7357-70c6-4b54-8e97-9860dd0618e7-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.527638 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njg9d\" (UniqueName: \"kubernetes.io/projected/319a7357-70c6-4b54-8e97-9860dd0618e7-kube-api-access-njg9d\") on node \"crc\" DevicePath \"\"" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.527723 4806 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.527824 4806 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/319a7357-70c6-4b54-8e97-9860dd0618e7-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.528128 4806 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.528196 4806 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/319a7357-70c6-4b54-8e97-9860dd0618e7-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.529068 4806 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.529257 4806 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/319a7357-70c6-4b54-8e97-9860dd0618e7-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.550331 4806 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.632225 4806 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.858997 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"319a7357-70c6-4b54-8e97-9860dd0618e7","Type":"ContainerDied","Data":"150f0528353b9920c10e0f675e8348ce6f78c30ca3ac89fc7ba1504ff7b89cf2"} Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.859059 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="150f0528353b9920c10e0f675e8348ce6f78c30ca3ac89fc7ba1504ff7b89cf2" Dec 04 05:12:38 crc kubenswrapper[4806]: I1204 05:12:38.859159 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 04 05:12:39 crc kubenswrapper[4806]: I1204 05:12:39.792395 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x5s92"] Dec 04 05:12:39 crc kubenswrapper[4806]: E1204 05:12:39.793125 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a583375-a2ee-469c-a6c3-0c3315fc1833" containerName="extract-utilities" Dec 04 05:12:39 crc kubenswrapper[4806]: I1204 05:12:39.793141 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a583375-a2ee-469c-a6c3-0c3315fc1833" containerName="extract-utilities" Dec 04 05:12:39 crc kubenswrapper[4806]: E1204 05:12:39.793157 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a583375-a2ee-469c-a6c3-0c3315fc1833" containerName="extract-content" Dec 04 05:12:39 crc kubenswrapper[4806]: I1204 05:12:39.793165 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a583375-a2ee-469c-a6c3-0c3315fc1833" containerName="extract-content" Dec 04 05:12:39 crc kubenswrapper[4806]: E1204 05:12:39.793183 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a583375-a2ee-469c-a6c3-0c3315fc1833" containerName="registry-server" Dec 04 05:12:39 crc kubenswrapper[4806]: I1204 05:12:39.793190 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a583375-a2ee-469c-a6c3-0c3315fc1833" containerName="registry-server" Dec 04 05:12:39 crc kubenswrapper[4806]: E1204 05:12:39.793208 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="319a7357-70c6-4b54-8e97-9860dd0618e7" containerName="tempest-tests-tempest-tests-runner" Dec 04 05:12:39 crc kubenswrapper[4806]: I1204 05:12:39.793217 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="319a7357-70c6-4b54-8e97-9860dd0618e7" containerName="tempest-tests-tempest-tests-runner" Dec 04 05:12:39 crc kubenswrapper[4806]: I1204 05:12:39.793445 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a583375-a2ee-469c-a6c3-0c3315fc1833" containerName="registry-server" Dec 04 05:12:39 crc kubenswrapper[4806]: I1204 05:12:39.793466 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="319a7357-70c6-4b54-8e97-9860dd0618e7" containerName="tempest-tests-tempest-tests-runner" Dec 04 05:12:39 crc kubenswrapper[4806]: I1204 05:12:39.795072 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:12:39 crc kubenswrapper[4806]: I1204 05:12:39.815326 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x5s92"] Dec 04 05:12:39 crc kubenswrapper[4806]: I1204 05:12:39.959978 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-catalog-content\") pod \"redhat-operators-x5s92\" (UID: \"a5582b1a-b88e-4b85-b7ec-a3c13695ea99\") " pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:12:39 crc kubenswrapper[4806]: I1204 05:12:39.960068 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-utilities\") pod \"redhat-operators-x5s92\" (UID: \"a5582b1a-b88e-4b85-b7ec-a3c13695ea99\") " pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:12:39 crc kubenswrapper[4806]: I1204 05:12:39.960111 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kr4r\" (UniqueName: \"kubernetes.io/projected/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-kube-api-access-5kr4r\") pod \"redhat-operators-x5s92\" (UID: \"a5582b1a-b88e-4b85-b7ec-a3c13695ea99\") " pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:12:40 crc kubenswrapper[4806]: I1204 05:12:40.062605 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-catalog-content\") pod \"redhat-operators-x5s92\" (UID: \"a5582b1a-b88e-4b85-b7ec-a3c13695ea99\") " pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:12:40 crc kubenswrapper[4806]: I1204 05:12:40.062728 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-utilities\") pod \"redhat-operators-x5s92\" (UID: \"a5582b1a-b88e-4b85-b7ec-a3c13695ea99\") " pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:12:40 crc kubenswrapper[4806]: I1204 05:12:40.062790 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kr4r\" (UniqueName: \"kubernetes.io/projected/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-kube-api-access-5kr4r\") pod \"redhat-operators-x5s92\" (UID: \"a5582b1a-b88e-4b85-b7ec-a3c13695ea99\") " pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:12:40 crc kubenswrapper[4806]: I1204 05:12:40.063845 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-catalog-content\") pod \"redhat-operators-x5s92\" (UID: \"a5582b1a-b88e-4b85-b7ec-a3c13695ea99\") " pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:12:40 crc kubenswrapper[4806]: I1204 05:12:40.063998 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-utilities\") pod \"redhat-operators-x5s92\" (UID: \"a5582b1a-b88e-4b85-b7ec-a3c13695ea99\") " pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:12:40 crc kubenswrapper[4806]: I1204 05:12:40.088999 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kr4r\" (UniqueName: \"kubernetes.io/projected/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-kube-api-access-5kr4r\") pod \"redhat-operators-x5s92\" (UID: \"a5582b1a-b88e-4b85-b7ec-a3c13695ea99\") " pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:12:40 crc kubenswrapper[4806]: I1204 05:12:40.120017 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:12:40 crc kubenswrapper[4806]: I1204 05:12:40.613336 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x5s92"] Dec 04 05:12:40 crc kubenswrapper[4806]: I1204 05:12:40.878851 4806 generic.go:334] "Generic (PLEG): container finished" podID="a5582b1a-b88e-4b85-b7ec-a3c13695ea99" containerID="46e8dfafef9666ddcb4086c6d2666fb4c2a7d72e120e30dee5bad2d78468dd14" exitCode=0 Dec 04 05:12:40 crc kubenswrapper[4806]: I1204 05:12:40.879259 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5s92" event={"ID":"a5582b1a-b88e-4b85-b7ec-a3c13695ea99","Type":"ContainerDied","Data":"46e8dfafef9666ddcb4086c6d2666fb4c2a7d72e120e30dee5bad2d78468dd14"} Dec 04 05:12:40 crc kubenswrapper[4806]: I1204 05:12:40.879291 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5s92" event={"ID":"a5582b1a-b88e-4b85-b7ec-a3c13695ea99","Type":"ContainerStarted","Data":"7cfd72c98388d274360a39d316d8f3f62421eef055644b2a22203fcf40235a6c"} Dec 04 05:12:40 crc kubenswrapper[4806]: I1204 05:12:40.882442 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 05:12:41 crc kubenswrapper[4806]: I1204 05:12:41.895588 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5s92" event={"ID":"a5582b1a-b88e-4b85-b7ec-a3c13695ea99","Type":"ContainerStarted","Data":"95826f18657e0801681835679391ae412745d9a57e82baedaf4bb6d19a411731"} Dec 04 05:12:44 crc kubenswrapper[4806]: I1204 05:12:44.921093 4806 generic.go:334] "Generic (PLEG): container finished" podID="a5582b1a-b88e-4b85-b7ec-a3c13695ea99" containerID="95826f18657e0801681835679391ae412745d9a57e82baedaf4bb6d19a411731" exitCode=0 Dec 04 05:12:44 crc kubenswrapper[4806]: I1204 05:12:44.921124 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5s92" event={"ID":"a5582b1a-b88e-4b85-b7ec-a3c13695ea99","Type":"ContainerDied","Data":"95826f18657e0801681835679391ae412745d9a57e82baedaf4bb6d19a411731"} Dec 04 05:12:45 crc kubenswrapper[4806]: I1204 05:12:45.938449 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5s92" event={"ID":"a5582b1a-b88e-4b85-b7ec-a3c13695ea99","Type":"ContainerStarted","Data":"7c48ff440613149464b55c7a77c2f8622ce5c7c534501bfde2a1dc392aeb1450"} Dec 04 05:12:45 crc kubenswrapper[4806]: I1204 05:12:45.962033 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x5s92" podStartSLOduration=2.448816264 podStartE2EDuration="6.962013837s" podCreationTimestamp="2025-12-04 05:12:39 +0000 UTC" firstStartedPulling="2025-12-04 05:12:40.882139295 +0000 UTC m=+4675.740652223" lastFinishedPulling="2025-12-04 05:12:45.395336848 +0000 UTC m=+4680.253849796" observedRunningTime="2025-12-04 05:12:45.956819994 +0000 UTC m=+4680.815332922" watchObservedRunningTime="2025-12-04 05:12:45.962013837 +0000 UTC m=+4680.820526765" Dec 04 05:12:49 crc kubenswrapper[4806]: I1204 05:12:49.608952 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 04 05:12:49 crc kubenswrapper[4806]: I1204 05:12:49.611829 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 05:12:49 crc kubenswrapper[4806]: I1204 05:12:49.616331 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-jb6sp" Dec 04 05:12:49 crc kubenswrapper[4806]: I1204 05:12:49.619170 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 04 05:12:49 crc kubenswrapper[4806]: I1204 05:12:49.753525 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"4922f2b6-05e3-4f78-a14a-46ca2668edfb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 05:12:49 crc kubenswrapper[4806]: I1204 05:12:49.753601 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57h7p\" (UniqueName: \"kubernetes.io/projected/4922f2b6-05e3-4f78-a14a-46ca2668edfb-kube-api-access-57h7p\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"4922f2b6-05e3-4f78-a14a-46ca2668edfb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 05:12:49 crc kubenswrapper[4806]: I1204 05:12:49.855627 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"4922f2b6-05e3-4f78-a14a-46ca2668edfb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 05:12:49 crc kubenswrapper[4806]: I1204 05:12:49.855714 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57h7p\" (UniqueName: \"kubernetes.io/projected/4922f2b6-05e3-4f78-a14a-46ca2668edfb-kube-api-access-57h7p\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"4922f2b6-05e3-4f78-a14a-46ca2668edfb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 05:12:49 crc kubenswrapper[4806]: I1204 05:12:49.857146 4806 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"4922f2b6-05e3-4f78-a14a-46ca2668edfb\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 05:12:49 crc kubenswrapper[4806]: I1204 05:12:49.888765 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57h7p\" (UniqueName: \"kubernetes.io/projected/4922f2b6-05e3-4f78-a14a-46ca2668edfb-kube-api-access-57h7p\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"4922f2b6-05e3-4f78-a14a-46ca2668edfb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 05:12:49 crc kubenswrapper[4806]: I1204 05:12:49.907017 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"4922f2b6-05e3-4f78-a14a-46ca2668edfb\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 05:12:49 crc kubenswrapper[4806]: I1204 05:12:49.936430 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 04 05:12:50 crc kubenswrapper[4806]: I1204 05:12:50.121125 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:12:50 crc kubenswrapper[4806]: I1204 05:12:50.121587 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:12:50 crc kubenswrapper[4806]: I1204 05:12:50.468045 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 04 05:12:51 crc kubenswrapper[4806]: I1204 05:12:51.018425 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"4922f2b6-05e3-4f78-a14a-46ca2668edfb","Type":"ContainerStarted","Data":"02aac31f8007954f877399d02afe34b9122c5af3ad1292a3a665e7bc6f7ab060"} Dec 04 05:12:51 crc kubenswrapper[4806]: I1204 05:12:51.178877 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x5s92" podUID="a5582b1a-b88e-4b85-b7ec-a3c13695ea99" containerName="registry-server" probeResult="failure" output=< Dec 04 05:12:51 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Dec 04 05:12:51 crc kubenswrapper[4806]: > Dec 04 05:12:52 crc kubenswrapper[4806]: I1204 05:12:52.027966 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"4922f2b6-05e3-4f78-a14a-46ca2668edfb","Type":"ContainerStarted","Data":"1d05c8af623278773d124e11c93740d58449af64209746f23b8528c921543735"} Dec 04 05:12:52 crc kubenswrapper[4806]: I1204 05:12:52.046889 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.193162441 podStartE2EDuration="3.046868255s" podCreationTimestamp="2025-12-04 05:12:49 +0000 UTC" firstStartedPulling="2025-12-04 05:12:50.47549886 +0000 UTC m=+4685.334011788" lastFinishedPulling="2025-12-04 05:12:51.329204674 +0000 UTC m=+4686.187717602" observedRunningTime="2025-12-04 05:12:52.040528786 +0000 UTC m=+4686.899041714" watchObservedRunningTime="2025-12-04 05:12:52.046868255 +0000 UTC m=+4686.905381183" Dec 04 05:12:57 crc kubenswrapper[4806]: I1204 05:12:57.047209 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:12:57 crc kubenswrapper[4806]: I1204 05:12:57.048478 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:13:00 crc kubenswrapper[4806]: I1204 05:13:00.164783 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:13:00 crc kubenswrapper[4806]: I1204 05:13:00.211194 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:13:00 crc kubenswrapper[4806]: I1204 05:13:00.410043 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x5s92"] Dec 04 05:13:02 crc kubenswrapper[4806]: I1204 05:13:02.123613 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x5s92" podUID="a5582b1a-b88e-4b85-b7ec-a3c13695ea99" containerName="registry-server" containerID="cri-o://7c48ff440613149464b55c7a77c2f8622ce5c7c534501bfde2a1dc392aeb1450" gracePeriod=2 Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.076264 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.134397 4806 generic.go:334] "Generic (PLEG): container finished" podID="a5582b1a-b88e-4b85-b7ec-a3c13695ea99" containerID="7c48ff440613149464b55c7a77c2f8622ce5c7c534501bfde2a1dc392aeb1450" exitCode=0 Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.134454 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5s92" event={"ID":"a5582b1a-b88e-4b85-b7ec-a3c13695ea99","Type":"ContainerDied","Data":"7c48ff440613149464b55c7a77c2f8622ce5c7c534501bfde2a1dc392aeb1450"} Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.134521 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x5s92" event={"ID":"a5582b1a-b88e-4b85-b7ec-a3c13695ea99","Type":"ContainerDied","Data":"7cfd72c98388d274360a39d316d8f3f62421eef055644b2a22203fcf40235a6c"} Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.134545 4806 scope.go:117] "RemoveContainer" containerID="7c48ff440613149464b55c7a77c2f8622ce5c7c534501bfde2a1dc392aeb1450" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.135790 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x5s92" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.176170 4806 scope.go:117] "RemoveContainer" containerID="95826f18657e0801681835679391ae412745d9a57e82baedaf4bb6d19a411731" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.197148 4806 scope.go:117] "RemoveContainer" containerID="46e8dfafef9666ddcb4086c6d2666fb4c2a7d72e120e30dee5bad2d78468dd14" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.221302 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-catalog-content\") pod \"a5582b1a-b88e-4b85-b7ec-a3c13695ea99\" (UID: \"a5582b1a-b88e-4b85-b7ec-a3c13695ea99\") " Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.221406 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kr4r\" (UniqueName: \"kubernetes.io/projected/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-kube-api-access-5kr4r\") pod \"a5582b1a-b88e-4b85-b7ec-a3c13695ea99\" (UID: \"a5582b1a-b88e-4b85-b7ec-a3c13695ea99\") " Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.221499 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-utilities\") pod \"a5582b1a-b88e-4b85-b7ec-a3c13695ea99\" (UID: \"a5582b1a-b88e-4b85-b7ec-a3c13695ea99\") " Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.222184 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-utilities" (OuterVolumeSpecName: "utilities") pod "a5582b1a-b88e-4b85-b7ec-a3c13695ea99" (UID: "a5582b1a-b88e-4b85-b7ec-a3c13695ea99"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.228422 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-kube-api-access-5kr4r" (OuterVolumeSpecName: "kube-api-access-5kr4r") pod "a5582b1a-b88e-4b85-b7ec-a3c13695ea99" (UID: "a5582b1a-b88e-4b85-b7ec-a3c13695ea99"). InnerVolumeSpecName "kube-api-access-5kr4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.303131 4806 scope.go:117] "RemoveContainer" containerID="7c48ff440613149464b55c7a77c2f8622ce5c7c534501bfde2a1dc392aeb1450" Dec 04 05:13:03 crc kubenswrapper[4806]: E1204 05:13:03.303576 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c48ff440613149464b55c7a77c2f8622ce5c7c534501bfde2a1dc392aeb1450\": container with ID starting with 7c48ff440613149464b55c7a77c2f8622ce5c7c534501bfde2a1dc392aeb1450 not found: ID does not exist" containerID="7c48ff440613149464b55c7a77c2f8622ce5c7c534501bfde2a1dc392aeb1450" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.303629 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c48ff440613149464b55c7a77c2f8622ce5c7c534501bfde2a1dc392aeb1450"} err="failed to get container status \"7c48ff440613149464b55c7a77c2f8622ce5c7c534501bfde2a1dc392aeb1450\": rpc error: code = NotFound desc = could not find container \"7c48ff440613149464b55c7a77c2f8622ce5c7c534501bfde2a1dc392aeb1450\": container with ID starting with 7c48ff440613149464b55c7a77c2f8622ce5c7c534501bfde2a1dc392aeb1450 not found: ID does not exist" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.303661 4806 scope.go:117] "RemoveContainer" containerID="95826f18657e0801681835679391ae412745d9a57e82baedaf4bb6d19a411731" Dec 04 05:13:03 crc kubenswrapper[4806]: E1204 05:13:03.304069 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95826f18657e0801681835679391ae412745d9a57e82baedaf4bb6d19a411731\": container with ID starting with 95826f18657e0801681835679391ae412745d9a57e82baedaf4bb6d19a411731 not found: ID does not exist" containerID="95826f18657e0801681835679391ae412745d9a57e82baedaf4bb6d19a411731" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.304102 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95826f18657e0801681835679391ae412745d9a57e82baedaf4bb6d19a411731"} err="failed to get container status \"95826f18657e0801681835679391ae412745d9a57e82baedaf4bb6d19a411731\": rpc error: code = NotFound desc = could not find container \"95826f18657e0801681835679391ae412745d9a57e82baedaf4bb6d19a411731\": container with ID starting with 95826f18657e0801681835679391ae412745d9a57e82baedaf4bb6d19a411731 not found: ID does not exist" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.304128 4806 scope.go:117] "RemoveContainer" containerID="46e8dfafef9666ddcb4086c6d2666fb4c2a7d72e120e30dee5bad2d78468dd14" Dec 04 05:13:03 crc kubenswrapper[4806]: E1204 05:13:03.304331 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46e8dfafef9666ddcb4086c6d2666fb4c2a7d72e120e30dee5bad2d78468dd14\": container with ID starting with 46e8dfafef9666ddcb4086c6d2666fb4c2a7d72e120e30dee5bad2d78468dd14 not found: ID does not exist" containerID="46e8dfafef9666ddcb4086c6d2666fb4c2a7d72e120e30dee5bad2d78468dd14" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.304354 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46e8dfafef9666ddcb4086c6d2666fb4c2a7d72e120e30dee5bad2d78468dd14"} err="failed to get container status \"46e8dfafef9666ddcb4086c6d2666fb4c2a7d72e120e30dee5bad2d78468dd14\": rpc error: code = NotFound desc = could not find container \"46e8dfafef9666ddcb4086c6d2666fb4c2a7d72e120e30dee5bad2d78468dd14\": container with ID starting with 46e8dfafef9666ddcb4086c6d2666fb4c2a7d72e120e30dee5bad2d78468dd14 not found: ID does not exist" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.323611 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kr4r\" (UniqueName: \"kubernetes.io/projected/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-kube-api-access-5kr4r\") on node \"crc\" DevicePath \"\"" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.323642 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.334363 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5582b1a-b88e-4b85-b7ec-a3c13695ea99" (UID: "a5582b1a-b88e-4b85-b7ec-a3c13695ea99"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.463533 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5582b1a-b88e-4b85-b7ec-a3c13695ea99-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.503060 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x5s92"] Dec 04 05:13:03 crc kubenswrapper[4806]: I1204 05:13:03.512242 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x5s92"] Dec 04 05:13:05 crc kubenswrapper[4806]: I1204 05:13:05.437027 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5582b1a-b88e-4b85-b7ec-a3c13695ea99" path="/var/lib/kubelet/pods/a5582b1a-b88e-4b85-b7ec-a3c13695ea99/volumes" Dec 04 05:13:14 crc kubenswrapper[4806]: I1204 05:13:14.773954 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zx5tm/must-gather-mxl86"] Dec 04 05:13:14 crc kubenswrapper[4806]: E1204 05:13:14.775082 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5582b1a-b88e-4b85-b7ec-a3c13695ea99" containerName="extract-content" Dec 04 05:13:14 crc kubenswrapper[4806]: I1204 05:13:14.775100 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5582b1a-b88e-4b85-b7ec-a3c13695ea99" containerName="extract-content" Dec 04 05:13:14 crc kubenswrapper[4806]: E1204 05:13:14.775122 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5582b1a-b88e-4b85-b7ec-a3c13695ea99" containerName="registry-server" Dec 04 05:13:14 crc kubenswrapper[4806]: I1204 05:13:14.775128 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5582b1a-b88e-4b85-b7ec-a3c13695ea99" containerName="registry-server" Dec 04 05:13:14 crc kubenswrapper[4806]: E1204 05:13:14.775151 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5582b1a-b88e-4b85-b7ec-a3c13695ea99" containerName="extract-utilities" Dec 04 05:13:14 crc kubenswrapper[4806]: I1204 05:13:14.775158 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5582b1a-b88e-4b85-b7ec-a3c13695ea99" containerName="extract-utilities" Dec 04 05:13:14 crc kubenswrapper[4806]: I1204 05:13:14.775401 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5582b1a-b88e-4b85-b7ec-a3c13695ea99" containerName="registry-server" Dec 04 05:13:14 crc kubenswrapper[4806]: I1204 05:13:14.776917 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx5tm/must-gather-mxl86" Dec 04 05:13:14 crc kubenswrapper[4806]: I1204 05:13:14.779285 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zx5tm"/"openshift-service-ca.crt" Dec 04 05:13:14 crc kubenswrapper[4806]: I1204 05:13:14.779445 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-zx5tm"/"default-dockercfg-47bnh" Dec 04 05:13:14 crc kubenswrapper[4806]: I1204 05:13:14.780683 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zx5tm"/"kube-root-ca.crt" Dec 04 05:13:14 crc kubenswrapper[4806]: I1204 05:13:14.794232 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zx5tm/must-gather-mxl86"] Dec 04 05:13:14 crc kubenswrapper[4806]: I1204 05:13:14.896463 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49-must-gather-output\") pod \"must-gather-mxl86\" (UID: \"1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49\") " pod="openshift-must-gather-zx5tm/must-gather-mxl86" Dec 04 05:13:14 crc kubenswrapper[4806]: I1204 05:13:14.896555 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkbvc\" (UniqueName: \"kubernetes.io/projected/1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49-kube-api-access-qkbvc\") pod \"must-gather-mxl86\" (UID: \"1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49\") " pod="openshift-must-gather-zx5tm/must-gather-mxl86" Dec 04 05:13:14 crc kubenswrapper[4806]: I1204 05:13:14.998080 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49-must-gather-output\") pod \"must-gather-mxl86\" (UID: \"1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49\") " pod="openshift-must-gather-zx5tm/must-gather-mxl86" Dec 04 05:13:14 crc kubenswrapper[4806]: I1204 05:13:14.998406 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkbvc\" (UniqueName: \"kubernetes.io/projected/1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49-kube-api-access-qkbvc\") pod \"must-gather-mxl86\" (UID: \"1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49\") " pod="openshift-must-gather-zx5tm/must-gather-mxl86" Dec 04 05:13:14 crc kubenswrapper[4806]: I1204 05:13:14.998655 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49-must-gather-output\") pod \"must-gather-mxl86\" (UID: \"1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49\") " pod="openshift-must-gather-zx5tm/must-gather-mxl86" Dec 04 05:13:15 crc kubenswrapper[4806]: I1204 05:13:15.017677 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkbvc\" (UniqueName: \"kubernetes.io/projected/1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49-kube-api-access-qkbvc\") pod \"must-gather-mxl86\" (UID: \"1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49\") " pod="openshift-must-gather-zx5tm/must-gather-mxl86" Dec 04 05:13:15 crc kubenswrapper[4806]: I1204 05:13:15.094896 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx5tm/must-gather-mxl86" Dec 04 05:13:15 crc kubenswrapper[4806]: I1204 05:13:15.675978 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zx5tm/must-gather-mxl86"] Dec 04 05:13:16 crc kubenswrapper[4806]: I1204 05:13:16.266867 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx5tm/must-gather-mxl86" event={"ID":"1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49","Type":"ContainerStarted","Data":"5835645015edda3b0c74eb5585d735f348a033fc9e6d11c6ab8cb1257cdca529"} Dec 04 05:13:21 crc kubenswrapper[4806]: I1204 05:13:21.325527 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx5tm/must-gather-mxl86" event={"ID":"1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49","Type":"ContainerStarted","Data":"a47a02f05a9c602c361615d3351a2df5e5eb547a262626b51245aab28658722d"} Dec 04 05:13:22 crc kubenswrapper[4806]: I1204 05:13:22.339137 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx5tm/must-gather-mxl86" event={"ID":"1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49","Type":"ContainerStarted","Data":"471251cb2cdde3685841d548b1587e8e8144feed36a229eea8646086651bfa91"} Dec 04 05:13:22 crc kubenswrapper[4806]: I1204 05:13:22.355323 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zx5tm/must-gather-mxl86" podStartSLOduration=3.853856638 podStartE2EDuration="8.355306992s" podCreationTimestamp="2025-12-04 05:13:14 +0000 UTC" firstStartedPulling="2025-12-04 05:13:15.676075793 +0000 UTC m=+4710.534588721" lastFinishedPulling="2025-12-04 05:13:20.177526147 +0000 UTC m=+4715.036039075" observedRunningTime="2025-12-04 05:13:22.354417295 +0000 UTC m=+4717.212930243" watchObservedRunningTime="2025-12-04 05:13:22.355306992 +0000 UTC m=+4717.213819920" Dec 04 05:13:26 crc kubenswrapper[4806]: I1204 05:13:26.343744 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zx5tm/crc-debug-rkm86"] Dec 04 05:13:26 crc kubenswrapper[4806]: I1204 05:13:26.345483 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx5tm/crc-debug-rkm86" Dec 04 05:13:26 crc kubenswrapper[4806]: I1204 05:13:26.436826 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7wvn\" (UniqueName: \"kubernetes.io/projected/ef3d5bd3-f929-47c0-80b0-9b33b3c029f8-kube-api-access-g7wvn\") pod \"crc-debug-rkm86\" (UID: \"ef3d5bd3-f929-47c0-80b0-9b33b3c029f8\") " pod="openshift-must-gather-zx5tm/crc-debug-rkm86" Dec 04 05:13:26 crc kubenswrapper[4806]: I1204 05:13:26.437211 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ef3d5bd3-f929-47c0-80b0-9b33b3c029f8-host\") pod \"crc-debug-rkm86\" (UID: \"ef3d5bd3-f929-47c0-80b0-9b33b3c029f8\") " pod="openshift-must-gather-zx5tm/crc-debug-rkm86" Dec 04 05:13:26 crc kubenswrapper[4806]: I1204 05:13:26.539345 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ef3d5bd3-f929-47c0-80b0-9b33b3c029f8-host\") pod \"crc-debug-rkm86\" (UID: \"ef3d5bd3-f929-47c0-80b0-9b33b3c029f8\") " pod="openshift-must-gather-zx5tm/crc-debug-rkm86" Dec 04 05:13:26 crc kubenswrapper[4806]: I1204 05:13:26.539590 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ef3d5bd3-f929-47c0-80b0-9b33b3c029f8-host\") pod \"crc-debug-rkm86\" (UID: \"ef3d5bd3-f929-47c0-80b0-9b33b3c029f8\") " pod="openshift-must-gather-zx5tm/crc-debug-rkm86" Dec 04 05:13:26 crc kubenswrapper[4806]: I1204 05:13:26.540033 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7wvn\" (UniqueName: \"kubernetes.io/projected/ef3d5bd3-f929-47c0-80b0-9b33b3c029f8-kube-api-access-g7wvn\") pod \"crc-debug-rkm86\" (UID: \"ef3d5bd3-f929-47c0-80b0-9b33b3c029f8\") " pod="openshift-must-gather-zx5tm/crc-debug-rkm86" Dec 04 05:13:26 crc kubenswrapper[4806]: I1204 05:13:26.563295 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7wvn\" (UniqueName: \"kubernetes.io/projected/ef3d5bd3-f929-47c0-80b0-9b33b3c029f8-kube-api-access-g7wvn\") pod \"crc-debug-rkm86\" (UID: \"ef3d5bd3-f929-47c0-80b0-9b33b3c029f8\") " pod="openshift-must-gather-zx5tm/crc-debug-rkm86" Dec 04 05:13:26 crc kubenswrapper[4806]: I1204 05:13:26.666857 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx5tm/crc-debug-rkm86" Dec 04 05:13:27 crc kubenswrapper[4806]: I1204 05:13:27.047691 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:13:27 crc kubenswrapper[4806]: I1204 05:13:27.047772 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:13:27 crc kubenswrapper[4806]: I1204 05:13:27.399533 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx5tm/crc-debug-rkm86" event={"ID":"ef3d5bd3-f929-47c0-80b0-9b33b3c029f8","Type":"ContainerStarted","Data":"ee91d4e4e0e48259999fa1b29d8d6d7c93b568929fbbdc9f5460203a9aed66f3"} Dec 04 05:13:38 crc kubenswrapper[4806]: I1204 05:13:38.521642 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx5tm/crc-debug-rkm86" event={"ID":"ef3d5bd3-f929-47c0-80b0-9b33b3c029f8","Type":"ContainerStarted","Data":"6d219f38466a04d25dd65d18ea472b0b0bf433de780216fe0326af4c64bc94d2"} Dec 04 05:13:38 crc kubenswrapper[4806]: I1204 05:13:38.541918 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zx5tm/crc-debug-rkm86" podStartSLOduration=1.912501712 podStartE2EDuration="12.54190082s" podCreationTimestamp="2025-12-04 05:13:26 +0000 UTC" firstStartedPulling="2025-12-04 05:13:26.713420645 +0000 UTC m=+4721.571933573" lastFinishedPulling="2025-12-04 05:13:37.342819753 +0000 UTC m=+4732.201332681" observedRunningTime="2025-12-04 05:13:38.533679031 +0000 UTC m=+4733.392191959" watchObservedRunningTime="2025-12-04 05:13:38.54190082 +0000 UTC m=+4733.400413748" Dec 04 05:13:57 crc kubenswrapper[4806]: I1204 05:13:57.047352 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:13:57 crc kubenswrapper[4806]: I1204 05:13:57.047873 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:13:57 crc kubenswrapper[4806]: I1204 05:13:57.047914 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 05:13:57 crc kubenswrapper[4806]: I1204 05:13:57.048557 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a1e6c4d9918f0691a954f6d7dd837700fe9fc560ddcdf3bc751c5b70d3e912e2"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 05:13:57 crc kubenswrapper[4806]: I1204 05:13:57.048606 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://a1e6c4d9918f0691a954f6d7dd837700fe9fc560ddcdf3bc751c5b70d3e912e2" gracePeriod=600 Dec 04 05:13:57 crc kubenswrapper[4806]: I1204 05:13:57.692736 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="a1e6c4d9918f0691a954f6d7dd837700fe9fc560ddcdf3bc751c5b70d3e912e2" exitCode=0 Dec 04 05:13:57 crc kubenswrapper[4806]: I1204 05:13:57.692797 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"a1e6c4d9918f0691a954f6d7dd837700fe9fc560ddcdf3bc751c5b70d3e912e2"} Dec 04 05:13:57 crc kubenswrapper[4806]: I1204 05:13:57.693114 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459"} Dec 04 05:13:57 crc kubenswrapper[4806]: I1204 05:13:57.693141 4806 scope.go:117] "RemoveContainer" containerID="aff149067c2df679f38649ac1d7dfe4cf435e838c714f355f74fc3082533e744" Dec 04 05:14:17 crc kubenswrapper[4806]: I1204 05:14:17.136357 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5q2vk"] Dec 04 05:14:17 crc kubenswrapper[4806]: I1204 05:14:17.138911 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:17 crc kubenswrapper[4806]: I1204 05:14:17.147844 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5q2vk"] Dec 04 05:14:17 crc kubenswrapper[4806]: I1204 05:14:17.260643 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5da2b6c-0475-4998-a93d-c44872cee325-catalog-content\") pod \"community-operators-5q2vk\" (UID: \"e5da2b6c-0475-4998-a93d-c44872cee325\") " pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:17 crc kubenswrapper[4806]: I1204 05:14:17.261064 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bgwb\" (UniqueName: \"kubernetes.io/projected/e5da2b6c-0475-4998-a93d-c44872cee325-kube-api-access-2bgwb\") pod \"community-operators-5q2vk\" (UID: \"e5da2b6c-0475-4998-a93d-c44872cee325\") " pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:17 crc kubenswrapper[4806]: I1204 05:14:17.261102 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5da2b6c-0475-4998-a93d-c44872cee325-utilities\") pod \"community-operators-5q2vk\" (UID: \"e5da2b6c-0475-4998-a93d-c44872cee325\") " pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:17 crc kubenswrapper[4806]: I1204 05:14:17.363559 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5da2b6c-0475-4998-a93d-c44872cee325-catalog-content\") pod \"community-operators-5q2vk\" (UID: \"e5da2b6c-0475-4998-a93d-c44872cee325\") " pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:17 crc kubenswrapper[4806]: I1204 05:14:17.363644 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bgwb\" (UniqueName: \"kubernetes.io/projected/e5da2b6c-0475-4998-a93d-c44872cee325-kube-api-access-2bgwb\") pod \"community-operators-5q2vk\" (UID: \"e5da2b6c-0475-4998-a93d-c44872cee325\") " pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:17 crc kubenswrapper[4806]: I1204 05:14:17.363677 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5da2b6c-0475-4998-a93d-c44872cee325-utilities\") pod \"community-operators-5q2vk\" (UID: \"e5da2b6c-0475-4998-a93d-c44872cee325\") " pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:17 crc kubenswrapper[4806]: I1204 05:14:17.364204 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5da2b6c-0475-4998-a93d-c44872cee325-utilities\") pod \"community-operators-5q2vk\" (UID: \"e5da2b6c-0475-4998-a93d-c44872cee325\") " pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:17 crc kubenswrapper[4806]: I1204 05:14:17.364459 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5da2b6c-0475-4998-a93d-c44872cee325-catalog-content\") pod \"community-operators-5q2vk\" (UID: \"e5da2b6c-0475-4998-a93d-c44872cee325\") " pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:17 crc kubenswrapper[4806]: I1204 05:14:17.394866 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bgwb\" (UniqueName: \"kubernetes.io/projected/e5da2b6c-0475-4998-a93d-c44872cee325-kube-api-access-2bgwb\") pod \"community-operators-5q2vk\" (UID: \"e5da2b6c-0475-4998-a93d-c44872cee325\") " pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:17 crc kubenswrapper[4806]: I1204 05:14:17.480626 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:18 crc kubenswrapper[4806]: W1204 05:14:18.086736 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5da2b6c_0475_4998_a93d_c44872cee325.slice/crio-addcf5f8a2c3f5948518119a4a10b1b11a0fe7a5e2dca2c4374310cf051b7953 WatchSource:0}: Error finding container addcf5f8a2c3f5948518119a4a10b1b11a0fe7a5e2dca2c4374310cf051b7953: Status 404 returned error can't find the container with id addcf5f8a2c3f5948518119a4a10b1b11a0fe7a5e2dca2c4374310cf051b7953 Dec 04 05:14:18 crc kubenswrapper[4806]: I1204 05:14:18.087282 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5q2vk"] Dec 04 05:14:18 crc kubenswrapper[4806]: I1204 05:14:18.875696 4806 generic.go:334] "Generic (PLEG): container finished" podID="e5da2b6c-0475-4998-a93d-c44872cee325" containerID="813e47f26876d31d94273d58d2377fec6669e137cf25d0d348b608f7e8f2e4bf" exitCode=0 Dec 04 05:14:18 crc kubenswrapper[4806]: I1204 05:14:18.875785 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5q2vk" event={"ID":"e5da2b6c-0475-4998-a93d-c44872cee325","Type":"ContainerDied","Data":"813e47f26876d31d94273d58d2377fec6669e137cf25d0d348b608f7e8f2e4bf"} Dec 04 05:14:18 crc kubenswrapper[4806]: I1204 05:14:18.878314 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5q2vk" event={"ID":"e5da2b6c-0475-4998-a93d-c44872cee325","Type":"ContainerStarted","Data":"addcf5f8a2c3f5948518119a4a10b1b11a0fe7a5e2dca2c4374310cf051b7953"} Dec 04 05:14:19 crc kubenswrapper[4806]: I1204 05:14:19.891431 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5q2vk" event={"ID":"e5da2b6c-0475-4998-a93d-c44872cee325","Type":"ContainerStarted","Data":"03059d9331ff887c68ff20387b582c0050219987d0c4e81a13c8ff3886a17071"} Dec 04 05:14:21 crc kubenswrapper[4806]: I1204 05:14:21.910782 4806 generic.go:334] "Generic (PLEG): container finished" podID="e5da2b6c-0475-4998-a93d-c44872cee325" containerID="03059d9331ff887c68ff20387b582c0050219987d0c4e81a13c8ff3886a17071" exitCode=0 Dec 04 05:14:21 crc kubenswrapper[4806]: I1204 05:14:21.910868 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5q2vk" event={"ID":"e5da2b6c-0475-4998-a93d-c44872cee325","Type":"ContainerDied","Data":"03059d9331ff887c68ff20387b582c0050219987d0c4e81a13c8ff3886a17071"} Dec 04 05:14:22 crc kubenswrapper[4806]: I1204 05:14:22.944530 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5q2vk" event={"ID":"e5da2b6c-0475-4998-a93d-c44872cee325","Type":"ContainerStarted","Data":"638aca48274d5b6bafbd25251b326a65409aa10b14ea0b288b99c7338c36556b"} Dec 04 05:14:22 crc kubenswrapper[4806]: I1204 05:14:22.983746 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5q2vk" podStartSLOduration=2.556567132 podStartE2EDuration="5.983720745s" podCreationTimestamp="2025-12-04 05:14:17 +0000 UTC" firstStartedPulling="2025-12-04 05:14:18.877565939 +0000 UTC m=+4773.736078867" lastFinishedPulling="2025-12-04 05:14:22.304719552 +0000 UTC m=+4777.163232480" observedRunningTime="2025-12-04 05:14:22.971003914 +0000 UTC m=+4777.829516852" watchObservedRunningTime="2025-12-04 05:14:22.983720745 +0000 UTC m=+4777.842233663" Dec 04 05:14:27 crc kubenswrapper[4806]: I1204 05:14:27.481016 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:27 crc kubenswrapper[4806]: I1204 05:14:27.482707 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:27 crc kubenswrapper[4806]: I1204 05:14:27.547053 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:28 crc kubenswrapper[4806]: I1204 05:14:28.050807 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:28 crc kubenswrapper[4806]: I1204 05:14:28.118094 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5q2vk"] Dec 04 05:14:30 crc kubenswrapper[4806]: I1204 05:14:30.009666 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5q2vk" podUID="e5da2b6c-0475-4998-a93d-c44872cee325" containerName="registry-server" containerID="cri-o://638aca48274d5b6bafbd25251b326a65409aa10b14ea0b288b99c7338c36556b" gracePeriod=2 Dec 04 05:14:30 crc kubenswrapper[4806]: I1204 05:14:30.564798 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:30 crc kubenswrapper[4806]: I1204 05:14:30.634533 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5da2b6c-0475-4998-a93d-c44872cee325-catalog-content\") pod \"e5da2b6c-0475-4998-a93d-c44872cee325\" (UID: \"e5da2b6c-0475-4998-a93d-c44872cee325\") " Dec 04 05:14:30 crc kubenswrapper[4806]: I1204 05:14:30.634599 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5da2b6c-0475-4998-a93d-c44872cee325-utilities\") pod \"e5da2b6c-0475-4998-a93d-c44872cee325\" (UID: \"e5da2b6c-0475-4998-a93d-c44872cee325\") " Dec 04 05:14:30 crc kubenswrapper[4806]: I1204 05:14:30.634762 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bgwb\" (UniqueName: \"kubernetes.io/projected/e5da2b6c-0475-4998-a93d-c44872cee325-kube-api-access-2bgwb\") pod \"e5da2b6c-0475-4998-a93d-c44872cee325\" (UID: \"e5da2b6c-0475-4998-a93d-c44872cee325\") " Dec 04 05:14:30 crc kubenswrapper[4806]: I1204 05:14:30.636638 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5da2b6c-0475-4998-a93d-c44872cee325-utilities" (OuterVolumeSpecName: "utilities") pod "e5da2b6c-0475-4998-a93d-c44872cee325" (UID: "e5da2b6c-0475-4998-a93d-c44872cee325"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:14:30 crc kubenswrapper[4806]: I1204 05:14:30.652655 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5da2b6c-0475-4998-a93d-c44872cee325-kube-api-access-2bgwb" (OuterVolumeSpecName: "kube-api-access-2bgwb") pod "e5da2b6c-0475-4998-a93d-c44872cee325" (UID: "e5da2b6c-0475-4998-a93d-c44872cee325"). InnerVolumeSpecName "kube-api-access-2bgwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:14:30 crc kubenswrapper[4806]: I1204 05:14:30.720803 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5da2b6c-0475-4998-a93d-c44872cee325-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5da2b6c-0475-4998-a93d-c44872cee325" (UID: "e5da2b6c-0475-4998-a93d-c44872cee325"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:14:30 crc kubenswrapper[4806]: I1204 05:14:30.737482 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5da2b6c-0475-4998-a93d-c44872cee325-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 05:14:30 crc kubenswrapper[4806]: I1204 05:14:30.737531 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5da2b6c-0475-4998-a93d-c44872cee325-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 05:14:30 crc kubenswrapper[4806]: I1204 05:14:30.737545 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bgwb\" (UniqueName: \"kubernetes.io/projected/e5da2b6c-0475-4998-a93d-c44872cee325-kube-api-access-2bgwb\") on node \"crc\" DevicePath \"\"" Dec 04 05:14:31 crc kubenswrapper[4806]: I1204 05:14:31.021770 4806 generic.go:334] "Generic (PLEG): container finished" podID="e5da2b6c-0475-4998-a93d-c44872cee325" containerID="638aca48274d5b6bafbd25251b326a65409aa10b14ea0b288b99c7338c36556b" exitCode=0 Dec 04 05:14:31 crc kubenswrapper[4806]: I1204 05:14:31.021814 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5q2vk" event={"ID":"e5da2b6c-0475-4998-a93d-c44872cee325","Type":"ContainerDied","Data":"638aca48274d5b6bafbd25251b326a65409aa10b14ea0b288b99c7338c36556b"} Dec 04 05:14:31 crc kubenswrapper[4806]: I1204 05:14:31.021843 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5q2vk" event={"ID":"e5da2b6c-0475-4998-a93d-c44872cee325","Type":"ContainerDied","Data":"addcf5f8a2c3f5948518119a4a10b1b11a0fe7a5e2dca2c4374310cf051b7953"} Dec 04 05:14:31 crc kubenswrapper[4806]: I1204 05:14:31.021860 4806 scope.go:117] "RemoveContainer" containerID="638aca48274d5b6bafbd25251b326a65409aa10b14ea0b288b99c7338c36556b" Dec 04 05:14:31 crc kubenswrapper[4806]: I1204 05:14:31.022009 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5q2vk" Dec 04 05:14:31 crc kubenswrapper[4806]: I1204 05:14:31.050278 4806 scope.go:117] "RemoveContainer" containerID="03059d9331ff887c68ff20387b582c0050219987d0c4e81a13c8ff3886a17071" Dec 04 05:14:31 crc kubenswrapper[4806]: I1204 05:14:31.064203 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5q2vk"] Dec 04 05:14:31 crc kubenswrapper[4806]: I1204 05:14:31.077313 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5q2vk"] Dec 04 05:14:31 crc kubenswrapper[4806]: I1204 05:14:31.080198 4806 scope.go:117] "RemoveContainer" containerID="813e47f26876d31d94273d58d2377fec6669e137cf25d0d348b608f7e8f2e4bf" Dec 04 05:14:31 crc kubenswrapper[4806]: I1204 05:14:31.161553 4806 scope.go:117] "RemoveContainer" containerID="638aca48274d5b6bafbd25251b326a65409aa10b14ea0b288b99c7338c36556b" Dec 04 05:14:31 crc kubenswrapper[4806]: E1204 05:14:31.165580 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"638aca48274d5b6bafbd25251b326a65409aa10b14ea0b288b99c7338c36556b\": container with ID starting with 638aca48274d5b6bafbd25251b326a65409aa10b14ea0b288b99c7338c36556b not found: ID does not exist" containerID="638aca48274d5b6bafbd25251b326a65409aa10b14ea0b288b99c7338c36556b" Dec 04 05:14:31 crc kubenswrapper[4806]: I1204 05:14:31.165633 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"638aca48274d5b6bafbd25251b326a65409aa10b14ea0b288b99c7338c36556b"} err="failed to get container status \"638aca48274d5b6bafbd25251b326a65409aa10b14ea0b288b99c7338c36556b\": rpc error: code = NotFound desc = could not find container \"638aca48274d5b6bafbd25251b326a65409aa10b14ea0b288b99c7338c36556b\": container with ID starting with 638aca48274d5b6bafbd25251b326a65409aa10b14ea0b288b99c7338c36556b not found: ID does not exist" Dec 04 05:14:31 crc kubenswrapper[4806]: I1204 05:14:31.165661 4806 scope.go:117] "RemoveContainer" containerID="03059d9331ff887c68ff20387b582c0050219987d0c4e81a13c8ff3886a17071" Dec 04 05:14:31 crc kubenswrapper[4806]: E1204 05:14:31.166037 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03059d9331ff887c68ff20387b582c0050219987d0c4e81a13c8ff3886a17071\": container with ID starting with 03059d9331ff887c68ff20387b582c0050219987d0c4e81a13c8ff3886a17071 not found: ID does not exist" containerID="03059d9331ff887c68ff20387b582c0050219987d0c4e81a13c8ff3886a17071" Dec 04 05:14:31 crc kubenswrapper[4806]: I1204 05:14:31.166064 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03059d9331ff887c68ff20387b582c0050219987d0c4e81a13c8ff3886a17071"} err="failed to get container status \"03059d9331ff887c68ff20387b582c0050219987d0c4e81a13c8ff3886a17071\": rpc error: code = NotFound desc = could not find container \"03059d9331ff887c68ff20387b582c0050219987d0c4e81a13c8ff3886a17071\": container with ID starting with 03059d9331ff887c68ff20387b582c0050219987d0c4e81a13c8ff3886a17071 not found: ID does not exist" Dec 04 05:14:31 crc kubenswrapper[4806]: I1204 05:14:31.166082 4806 scope.go:117] "RemoveContainer" containerID="813e47f26876d31d94273d58d2377fec6669e137cf25d0d348b608f7e8f2e4bf" Dec 04 05:14:31 crc kubenswrapper[4806]: E1204 05:14:31.166338 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"813e47f26876d31d94273d58d2377fec6669e137cf25d0d348b608f7e8f2e4bf\": container with ID starting with 813e47f26876d31d94273d58d2377fec6669e137cf25d0d348b608f7e8f2e4bf not found: ID does not exist" containerID="813e47f26876d31d94273d58d2377fec6669e137cf25d0d348b608f7e8f2e4bf" Dec 04 05:14:31 crc kubenswrapper[4806]: I1204 05:14:31.166363 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"813e47f26876d31d94273d58d2377fec6669e137cf25d0d348b608f7e8f2e4bf"} err="failed to get container status \"813e47f26876d31d94273d58d2377fec6669e137cf25d0d348b608f7e8f2e4bf\": rpc error: code = NotFound desc = could not find container \"813e47f26876d31d94273d58d2377fec6669e137cf25d0d348b608f7e8f2e4bf\": container with ID starting with 813e47f26876d31d94273d58d2377fec6669e137cf25d0d348b608f7e8f2e4bf not found: ID does not exist" Dec 04 05:14:31 crc kubenswrapper[4806]: I1204 05:14:31.436989 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5da2b6c-0475-4998-a93d-c44872cee325" path="/var/lib/kubelet/pods/e5da2b6c-0475-4998-a93d-c44872cee325/volumes" Dec 04 05:14:32 crc kubenswrapper[4806]: I1204 05:14:32.033409 4806 generic.go:334] "Generic (PLEG): container finished" podID="ef3d5bd3-f929-47c0-80b0-9b33b3c029f8" containerID="6d219f38466a04d25dd65d18ea472b0b0bf433de780216fe0326af4c64bc94d2" exitCode=0 Dec 04 05:14:32 crc kubenswrapper[4806]: I1204 05:14:32.033493 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx5tm/crc-debug-rkm86" event={"ID":"ef3d5bd3-f929-47c0-80b0-9b33b3c029f8","Type":"ContainerDied","Data":"6d219f38466a04d25dd65d18ea472b0b0bf433de780216fe0326af4c64bc94d2"} Dec 04 05:14:33 crc kubenswrapper[4806]: I1204 05:14:33.141800 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx5tm/crc-debug-rkm86" Dec 04 05:14:33 crc kubenswrapper[4806]: I1204 05:14:33.176644 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zx5tm/crc-debug-rkm86"] Dec 04 05:14:33 crc kubenswrapper[4806]: I1204 05:14:33.188511 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zx5tm/crc-debug-rkm86"] Dec 04 05:14:33 crc kubenswrapper[4806]: I1204 05:14:33.322823 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7wvn\" (UniqueName: \"kubernetes.io/projected/ef3d5bd3-f929-47c0-80b0-9b33b3c029f8-kube-api-access-g7wvn\") pod \"ef3d5bd3-f929-47c0-80b0-9b33b3c029f8\" (UID: \"ef3d5bd3-f929-47c0-80b0-9b33b3c029f8\") " Dec 04 05:14:33 crc kubenswrapper[4806]: I1204 05:14:33.322895 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ef3d5bd3-f929-47c0-80b0-9b33b3c029f8-host\") pod \"ef3d5bd3-f929-47c0-80b0-9b33b3c029f8\" (UID: \"ef3d5bd3-f929-47c0-80b0-9b33b3c029f8\") " Dec 04 05:14:33 crc kubenswrapper[4806]: I1204 05:14:33.323072 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ef3d5bd3-f929-47c0-80b0-9b33b3c029f8-host" (OuterVolumeSpecName: "host") pod "ef3d5bd3-f929-47c0-80b0-9b33b3c029f8" (UID: "ef3d5bd3-f929-47c0-80b0-9b33b3c029f8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 05:14:33 crc kubenswrapper[4806]: I1204 05:14:33.323669 4806 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ef3d5bd3-f929-47c0-80b0-9b33b3c029f8-host\") on node \"crc\" DevicePath \"\"" Dec 04 05:14:33 crc kubenswrapper[4806]: I1204 05:14:33.331037 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef3d5bd3-f929-47c0-80b0-9b33b3c029f8-kube-api-access-g7wvn" (OuterVolumeSpecName: "kube-api-access-g7wvn") pod "ef3d5bd3-f929-47c0-80b0-9b33b3c029f8" (UID: "ef3d5bd3-f929-47c0-80b0-9b33b3c029f8"). InnerVolumeSpecName "kube-api-access-g7wvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:14:33 crc kubenswrapper[4806]: I1204 05:14:33.425242 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7wvn\" (UniqueName: \"kubernetes.io/projected/ef3d5bd3-f929-47c0-80b0-9b33b3c029f8-kube-api-access-g7wvn\") on node \"crc\" DevicePath \"\"" Dec 04 05:14:33 crc kubenswrapper[4806]: I1204 05:14:33.440979 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef3d5bd3-f929-47c0-80b0-9b33b3c029f8" path="/var/lib/kubelet/pods/ef3d5bd3-f929-47c0-80b0-9b33b3c029f8/volumes" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.055275 4806 scope.go:117] "RemoveContainer" containerID="6d219f38466a04d25dd65d18ea472b0b0bf433de780216fe0326af4c64bc94d2" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.055317 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx5tm/crc-debug-rkm86" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.512363 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zx5tm/crc-debug-l54t6"] Dec 04 05:14:34 crc kubenswrapper[4806]: E1204 05:14:34.513117 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5da2b6c-0475-4998-a93d-c44872cee325" containerName="registry-server" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.513140 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5da2b6c-0475-4998-a93d-c44872cee325" containerName="registry-server" Dec 04 05:14:34 crc kubenswrapper[4806]: E1204 05:14:34.513191 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5da2b6c-0475-4998-a93d-c44872cee325" containerName="extract-utilities" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.513201 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5da2b6c-0475-4998-a93d-c44872cee325" containerName="extract-utilities" Dec 04 05:14:34 crc kubenswrapper[4806]: E1204 05:14:34.513228 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5da2b6c-0475-4998-a93d-c44872cee325" containerName="extract-content" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.513237 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5da2b6c-0475-4998-a93d-c44872cee325" containerName="extract-content" Dec 04 05:14:34 crc kubenswrapper[4806]: E1204 05:14:34.513275 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef3d5bd3-f929-47c0-80b0-9b33b3c029f8" containerName="container-00" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.513285 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef3d5bd3-f929-47c0-80b0-9b33b3c029f8" containerName="container-00" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.513557 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef3d5bd3-f929-47c0-80b0-9b33b3c029f8" containerName="container-00" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.513606 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5da2b6c-0475-4998-a93d-c44872cee325" containerName="registry-server" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.515487 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx5tm/crc-debug-l54t6" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.654341 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmk86\" (UniqueName: \"kubernetes.io/projected/e501f524-9f2c-4a24-919b-db93e864523d-kube-api-access-pmk86\") pod \"crc-debug-l54t6\" (UID: \"e501f524-9f2c-4a24-919b-db93e864523d\") " pod="openshift-must-gather-zx5tm/crc-debug-l54t6" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.654869 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e501f524-9f2c-4a24-919b-db93e864523d-host\") pod \"crc-debug-l54t6\" (UID: \"e501f524-9f2c-4a24-919b-db93e864523d\") " pod="openshift-must-gather-zx5tm/crc-debug-l54t6" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.756245 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmk86\" (UniqueName: \"kubernetes.io/projected/e501f524-9f2c-4a24-919b-db93e864523d-kube-api-access-pmk86\") pod \"crc-debug-l54t6\" (UID: \"e501f524-9f2c-4a24-919b-db93e864523d\") " pod="openshift-must-gather-zx5tm/crc-debug-l54t6" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.756497 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e501f524-9f2c-4a24-919b-db93e864523d-host\") pod \"crc-debug-l54t6\" (UID: \"e501f524-9f2c-4a24-919b-db93e864523d\") " pod="openshift-must-gather-zx5tm/crc-debug-l54t6" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.756613 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e501f524-9f2c-4a24-919b-db93e864523d-host\") pod \"crc-debug-l54t6\" (UID: \"e501f524-9f2c-4a24-919b-db93e864523d\") " pod="openshift-must-gather-zx5tm/crc-debug-l54t6" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.778755 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmk86\" (UniqueName: \"kubernetes.io/projected/e501f524-9f2c-4a24-919b-db93e864523d-kube-api-access-pmk86\") pod \"crc-debug-l54t6\" (UID: \"e501f524-9f2c-4a24-919b-db93e864523d\") " pod="openshift-must-gather-zx5tm/crc-debug-l54t6" Dec 04 05:14:34 crc kubenswrapper[4806]: I1204 05:14:34.841565 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx5tm/crc-debug-l54t6" Dec 04 05:14:35 crc kubenswrapper[4806]: I1204 05:14:35.064657 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx5tm/crc-debug-l54t6" event={"ID":"e501f524-9f2c-4a24-919b-db93e864523d","Type":"ContainerStarted","Data":"95718f753d3fbc0f35f89809719130a31152b5ebc48016cd84e77b532fb511ac"} Dec 04 05:14:36 crc kubenswrapper[4806]: I1204 05:14:36.081167 4806 generic.go:334] "Generic (PLEG): container finished" podID="e501f524-9f2c-4a24-919b-db93e864523d" containerID="333e066ce103c754b931cd382b76646a6ae9b277085c829f90d63556845c6290" exitCode=0 Dec 04 05:14:36 crc kubenswrapper[4806]: I1204 05:14:36.081207 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx5tm/crc-debug-l54t6" event={"ID":"e501f524-9f2c-4a24-919b-db93e864523d","Type":"ContainerDied","Data":"333e066ce103c754b931cd382b76646a6ae9b277085c829f90d63556845c6290"} Dec 04 05:14:37 crc kubenswrapper[4806]: I1204 05:14:37.187400 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx5tm/crc-debug-l54t6" Dec 04 05:14:37 crc kubenswrapper[4806]: I1204 05:14:37.203061 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e501f524-9f2c-4a24-919b-db93e864523d-host\") pod \"e501f524-9f2c-4a24-919b-db93e864523d\" (UID: \"e501f524-9f2c-4a24-919b-db93e864523d\") " Dec 04 05:14:37 crc kubenswrapper[4806]: I1204 05:14:37.203124 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmk86\" (UniqueName: \"kubernetes.io/projected/e501f524-9f2c-4a24-919b-db93e864523d-kube-api-access-pmk86\") pod \"e501f524-9f2c-4a24-919b-db93e864523d\" (UID: \"e501f524-9f2c-4a24-919b-db93e864523d\") " Dec 04 05:14:37 crc kubenswrapper[4806]: I1204 05:14:37.206077 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e501f524-9f2c-4a24-919b-db93e864523d-host" (OuterVolumeSpecName: "host") pod "e501f524-9f2c-4a24-919b-db93e864523d" (UID: "e501f524-9f2c-4a24-919b-db93e864523d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 05:14:37 crc kubenswrapper[4806]: I1204 05:14:37.228247 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e501f524-9f2c-4a24-919b-db93e864523d-kube-api-access-pmk86" (OuterVolumeSpecName: "kube-api-access-pmk86") pod "e501f524-9f2c-4a24-919b-db93e864523d" (UID: "e501f524-9f2c-4a24-919b-db93e864523d"). InnerVolumeSpecName "kube-api-access-pmk86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:14:37 crc kubenswrapper[4806]: I1204 05:14:37.304339 4806 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e501f524-9f2c-4a24-919b-db93e864523d-host\") on node \"crc\" DevicePath \"\"" Dec 04 05:14:37 crc kubenswrapper[4806]: I1204 05:14:37.304368 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmk86\" (UniqueName: \"kubernetes.io/projected/e501f524-9f2c-4a24-919b-db93e864523d-kube-api-access-pmk86\") on node \"crc\" DevicePath \"\"" Dec 04 05:14:38 crc kubenswrapper[4806]: I1204 05:14:38.106831 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx5tm/crc-debug-l54t6" event={"ID":"e501f524-9f2c-4a24-919b-db93e864523d","Type":"ContainerDied","Data":"95718f753d3fbc0f35f89809719130a31152b5ebc48016cd84e77b532fb511ac"} Dec 04 05:14:38 crc kubenswrapper[4806]: I1204 05:14:38.107204 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95718f753d3fbc0f35f89809719130a31152b5ebc48016cd84e77b532fb511ac" Dec 04 05:14:38 crc kubenswrapper[4806]: I1204 05:14:38.106963 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx5tm/crc-debug-l54t6" Dec 04 05:14:38 crc kubenswrapper[4806]: I1204 05:14:38.189398 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zx5tm/crc-debug-l54t6"] Dec 04 05:14:38 crc kubenswrapper[4806]: I1204 05:14:38.193440 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zx5tm/crc-debug-l54t6"] Dec 04 05:14:39 crc kubenswrapper[4806]: I1204 05:14:39.361657 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zx5tm/crc-debug-9j964"] Dec 04 05:14:39 crc kubenswrapper[4806]: E1204 05:14:39.362382 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e501f524-9f2c-4a24-919b-db93e864523d" containerName="container-00" Dec 04 05:14:39 crc kubenswrapper[4806]: I1204 05:14:39.362397 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="e501f524-9f2c-4a24-919b-db93e864523d" containerName="container-00" Dec 04 05:14:39 crc kubenswrapper[4806]: I1204 05:14:39.362609 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="e501f524-9f2c-4a24-919b-db93e864523d" containerName="container-00" Dec 04 05:14:39 crc kubenswrapper[4806]: I1204 05:14:39.363278 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx5tm/crc-debug-9j964" Dec 04 05:14:39 crc kubenswrapper[4806]: I1204 05:14:39.435335 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e501f524-9f2c-4a24-919b-db93e864523d" path="/var/lib/kubelet/pods/e501f524-9f2c-4a24-919b-db93e864523d/volumes" Dec 04 05:14:39 crc kubenswrapper[4806]: I1204 05:14:39.538207 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a-host\") pod \"crc-debug-9j964\" (UID: \"7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a\") " pod="openshift-must-gather-zx5tm/crc-debug-9j964" Dec 04 05:14:39 crc kubenswrapper[4806]: I1204 05:14:39.538258 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxzw2\" (UniqueName: \"kubernetes.io/projected/7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a-kube-api-access-lxzw2\") pod \"crc-debug-9j964\" (UID: \"7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a\") " pod="openshift-must-gather-zx5tm/crc-debug-9j964" Dec 04 05:14:39 crc kubenswrapper[4806]: I1204 05:14:39.640725 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxzw2\" (UniqueName: \"kubernetes.io/projected/7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a-kube-api-access-lxzw2\") pod \"crc-debug-9j964\" (UID: \"7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a\") " pod="openshift-must-gather-zx5tm/crc-debug-9j964" Dec 04 05:14:39 crc kubenswrapper[4806]: I1204 05:14:39.640937 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a-host\") pod \"crc-debug-9j964\" (UID: \"7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a\") " pod="openshift-must-gather-zx5tm/crc-debug-9j964" Dec 04 05:14:39 crc kubenswrapper[4806]: I1204 05:14:39.641028 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a-host\") pod \"crc-debug-9j964\" (UID: \"7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a\") " pod="openshift-must-gather-zx5tm/crc-debug-9j964" Dec 04 05:14:39 crc kubenswrapper[4806]: I1204 05:14:39.661023 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxzw2\" (UniqueName: \"kubernetes.io/projected/7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a-kube-api-access-lxzw2\") pod \"crc-debug-9j964\" (UID: \"7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a\") " pod="openshift-must-gather-zx5tm/crc-debug-9j964" Dec 04 05:14:39 crc kubenswrapper[4806]: I1204 05:14:39.679105 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx5tm/crc-debug-9j964" Dec 04 05:14:39 crc kubenswrapper[4806]: W1204 05:14:39.709293 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fe53b3e_ccc8_4d30_b1f1_88b83bc3245a.slice/crio-e490a90fa8b6a105ff92cc99ca7581c08f11524faf6f7545f479703e7c190fb1 WatchSource:0}: Error finding container e490a90fa8b6a105ff92cc99ca7581c08f11524faf6f7545f479703e7c190fb1: Status 404 returned error can't find the container with id e490a90fa8b6a105ff92cc99ca7581c08f11524faf6f7545f479703e7c190fb1 Dec 04 05:14:40 crc kubenswrapper[4806]: I1204 05:14:40.133133 4806 generic.go:334] "Generic (PLEG): container finished" podID="7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a" containerID="6f3c9a18a419b4b35a0f92eb5b803a4a28a6f5806018ae80dda10ff2a98afa03" exitCode=0 Dec 04 05:14:40 crc kubenswrapper[4806]: I1204 05:14:40.133214 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx5tm/crc-debug-9j964" event={"ID":"7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a","Type":"ContainerDied","Data":"6f3c9a18a419b4b35a0f92eb5b803a4a28a6f5806018ae80dda10ff2a98afa03"} Dec 04 05:14:40 crc kubenswrapper[4806]: I1204 05:14:40.133786 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx5tm/crc-debug-9j964" event={"ID":"7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a","Type":"ContainerStarted","Data":"e490a90fa8b6a105ff92cc99ca7581c08f11524faf6f7545f479703e7c190fb1"} Dec 04 05:14:40 crc kubenswrapper[4806]: I1204 05:14:40.179375 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zx5tm/crc-debug-9j964"] Dec 04 05:14:40 crc kubenswrapper[4806]: I1204 05:14:40.190953 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zx5tm/crc-debug-9j964"] Dec 04 05:14:41 crc kubenswrapper[4806]: I1204 05:14:41.364233 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx5tm/crc-debug-9j964" Dec 04 05:14:41 crc kubenswrapper[4806]: I1204 05:14:41.476485 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxzw2\" (UniqueName: \"kubernetes.io/projected/7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a-kube-api-access-lxzw2\") pod \"7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a\" (UID: \"7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a\") " Dec 04 05:14:41 crc kubenswrapper[4806]: I1204 05:14:41.476764 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a-host\") pod \"7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a\" (UID: \"7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a\") " Dec 04 05:14:41 crc kubenswrapper[4806]: I1204 05:14:41.476905 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a-host" (OuterVolumeSpecName: "host") pod "7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a" (UID: "7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 05:14:41 crc kubenswrapper[4806]: I1204 05:14:41.477368 4806 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a-host\") on node \"crc\" DevicePath \"\"" Dec 04 05:14:41 crc kubenswrapper[4806]: I1204 05:14:41.496620 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a-kube-api-access-lxzw2" (OuterVolumeSpecName: "kube-api-access-lxzw2") pod "7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a" (UID: "7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a"). InnerVolumeSpecName "kube-api-access-lxzw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:14:41 crc kubenswrapper[4806]: I1204 05:14:41.579824 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxzw2\" (UniqueName: \"kubernetes.io/projected/7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a-kube-api-access-lxzw2\") on node \"crc\" DevicePath \"\"" Dec 04 05:14:42 crc kubenswrapper[4806]: I1204 05:14:42.262015 4806 scope.go:117] "RemoveContainer" containerID="6f3c9a18a419b4b35a0f92eb5b803a4a28a6f5806018ae80dda10ff2a98afa03" Dec 04 05:14:42 crc kubenswrapper[4806]: I1204 05:14:42.262020 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx5tm/crc-debug-9j964" Dec 04 05:14:43 crc kubenswrapper[4806]: I1204 05:14:43.454812 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a" path="/var/lib/kubelet/pods/7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a/volumes" Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.149700 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf"] Dec 04 05:15:00 crc kubenswrapper[4806]: E1204 05:15:00.150687 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a" containerName="container-00" Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.150707 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a" containerName="container-00" Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.150987 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fe53b3e-ccc8-4d30-b1f1-88b83bc3245a" containerName="container-00" Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.151764 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf" Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.162623 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.162668 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.189200 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf"] Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.239693 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c37fbb2-7a56-413d-a7a3-67c6adc26608-secret-volume\") pod \"collect-profiles-29413755-kvvwf\" (UID: \"6c37fbb2-7a56-413d-a7a3-67c6adc26608\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf" Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.239749 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c37fbb2-7a56-413d-a7a3-67c6adc26608-config-volume\") pod \"collect-profiles-29413755-kvvwf\" (UID: \"6c37fbb2-7a56-413d-a7a3-67c6adc26608\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf" Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.239822 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm6fc\" (UniqueName: \"kubernetes.io/projected/6c37fbb2-7a56-413d-a7a3-67c6adc26608-kube-api-access-hm6fc\") pod \"collect-profiles-29413755-kvvwf\" (UID: \"6c37fbb2-7a56-413d-a7a3-67c6adc26608\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf" Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.341628 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c37fbb2-7a56-413d-a7a3-67c6adc26608-secret-volume\") pod \"collect-profiles-29413755-kvvwf\" (UID: \"6c37fbb2-7a56-413d-a7a3-67c6adc26608\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf" Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.341682 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c37fbb2-7a56-413d-a7a3-67c6adc26608-config-volume\") pod \"collect-profiles-29413755-kvvwf\" (UID: \"6c37fbb2-7a56-413d-a7a3-67c6adc26608\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf" Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.341734 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm6fc\" (UniqueName: \"kubernetes.io/projected/6c37fbb2-7a56-413d-a7a3-67c6adc26608-kube-api-access-hm6fc\") pod \"collect-profiles-29413755-kvvwf\" (UID: \"6c37fbb2-7a56-413d-a7a3-67c6adc26608\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf" Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.342698 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c37fbb2-7a56-413d-a7a3-67c6adc26608-config-volume\") pod \"collect-profiles-29413755-kvvwf\" (UID: \"6c37fbb2-7a56-413d-a7a3-67c6adc26608\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf" Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.348019 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c37fbb2-7a56-413d-a7a3-67c6adc26608-secret-volume\") pod \"collect-profiles-29413755-kvvwf\" (UID: \"6c37fbb2-7a56-413d-a7a3-67c6adc26608\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf" Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.361812 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm6fc\" (UniqueName: \"kubernetes.io/projected/6c37fbb2-7a56-413d-a7a3-67c6adc26608-kube-api-access-hm6fc\") pod \"collect-profiles-29413755-kvvwf\" (UID: \"6c37fbb2-7a56-413d-a7a3-67c6adc26608\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf" Dec 04 05:15:00 crc kubenswrapper[4806]: I1204 05:15:00.521635 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf" Dec 04 05:15:01 crc kubenswrapper[4806]: I1204 05:15:01.016131 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf"] Dec 04 05:15:01 crc kubenswrapper[4806]: I1204 05:15:01.422802 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf" event={"ID":"6c37fbb2-7a56-413d-a7a3-67c6adc26608","Type":"ContainerStarted","Data":"a6600263e11468b01ada1837c41d6ecc5b8fb502eec0bf47acccb677988e2664"} Dec 04 05:15:02 crc kubenswrapper[4806]: I1204 05:15:02.442906 4806 generic.go:334] "Generic (PLEG): container finished" podID="6c37fbb2-7a56-413d-a7a3-67c6adc26608" containerID="c1ade035a9b9ebb97d8ce8f166490115c0d1ee986cfea204a00f87e396206e33" exitCode=0 Dec 04 05:15:02 crc kubenswrapper[4806]: I1204 05:15:02.443278 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf" event={"ID":"6c37fbb2-7a56-413d-a7a3-67c6adc26608","Type":"ContainerDied","Data":"c1ade035a9b9ebb97d8ce8f166490115c0d1ee986cfea204a00f87e396206e33"} Dec 04 05:15:03 crc kubenswrapper[4806]: I1204 05:15:03.830621 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf" Dec 04 05:15:03 crc kubenswrapper[4806]: I1204 05:15:03.924679 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c37fbb2-7a56-413d-a7a3-67c6adc26608-secret-volume\") pod \"6c37fbb2-7a56-413d-a7a3-67c6adc26608\" (UID: \"6c37fbb2-7a56-413d-a7a3-67c6adc26608\") " Dec 04 05:15:03 crc kubenswrapper[4806]: I1204 05:15:03.925091 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c37fbb2-7a56-413d-a7a3-67c6adc26608-config-volume\") pod \"6c37fbb2-7a56-413d-a7a3-67c6adc26608\" (UID: \"6c37fbb2-7a56-413d-a7a3-67c6adc26608\") " Dec 04 05:15:03 crc kubenswrapper[4806]: I1204 05:15:03.925234 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hm6fc\" (UniqueName: \"kubernetes.io/projected/6c37fbb2-7a56-413d-a7a3-67c6adc26608-kube-api-access-hm6fc\") pod \"6c37fbb2-7a56-413d-a7a3-67c6adc26608\" (UID: \"6c37fbb2-7a56-413d-a7a3-67c6adc26608\") " Dec 04 05:15:03 crc kubenswrapper[4806]: I1204 05:15:03.926532 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c37fbb2-7a56-413d-a7a3-67c6adc26608-config-volume" (OuterVolumeSpecName: "config-volume") pod "6c37fbb2-7a56-413d-a7a3-67c6adc26608" (UID: "6c37fbb2-7a56-413d-a7a3-67c6adc26608"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 05:15:03 crc kubenswrapper[4806]: I1204 05:15:03.932897 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c37fbb2-7a56-413d-a7a3-67c6adc26608-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6c37fbb2-7a56-413d-a7a3-67c6adc26608" (UID: "6c37fbb2-7a56-413d-a7a3-67c6adc26608"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 05:15:03 crc kubenswrapper[4806]: I1204 05:15:03.933571 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c37fbb2-7a56-413d-a7a3-67c6adc26608-kube-api-access-hm6fc" (OuterVolumeSpecName: "kube-api-access-hm6fc") pod "6c37fbb2-7a56-413d-a7a3-67c6adc26608" (UID: "6c37fbb2-7a56-413d-a7a3-67c6adc26608"). InnerVolumeSpecName "kube-api-access-hm6fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:15:04 crc kubenswrapper[4806]: I1204 05:15:04.028001 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hm6fc\" (UniqueName: \"kubernetes.io/projected/6c37fbb2-7a56-413d-a7a3-67c6adc26608-kube-api-access-hm6fc\") on node \"crc\" DevicePath \"\"" Dec 04 05:15:04 crc kubenswrapper[4806]: I1204 05:15:04.028036 4806 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6c37fbb2-7a56-413d-a7a3-67c6adc26608-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 05:15:04 crc kubenswrapper[4806]: I1204 05:15:04.028047 4806 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6c37fbb2-7a56-413d-a7a3-67c6adc26608-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 05:15:04 crc kubenswrapper[4806]: I1204 05:15:04.461299 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf" event={"ID":"6c37fbb2-7a56-413d-a7a3-67c6adc26608","Type":"ContainerDied","Data":"a6600263e11468b01ada1837c41d6ecc5b8fb502eec0bf47acccb677988e2664"} Dec 04 05:15:04 crc kubenswrapper[4806]: I1204 05:15:04.461361 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6600263e11468b01ada1837c41d6ecc5b8fb502eec0bf47acccb677988e2664" Dec 04 05:15:04 crc kubenswrapper[4806]: I1204 05:15:04.461444 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413755-kvvwf" Dec 04 05:15:04 crc kubenswrapper[4806]: I1204 05:15:04.912230 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp"] Dec 04 05:15:04 crc kubenswrapper[4806]: I1204 05:15:04.921260 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413710-xt7pp"] Dec 04 05:15:05 crc kubenswrapper[4806]: I1204 05:15:05.438129 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e456fba1-e611-4f88-9ead-22a8c3c1bc0b" path="/var/lib/kubelet/pods/e456fba1-e611-4f88-9ead-22a8c3c1bc0b/volumes" Dec 04 05:15:11 crc kubenswrapper[4806]: I1204 05:15:11.901097 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-949b6795d-z7dx2_ef61aa0b-2b2c-481f-b22d-4ea770a91711/barbican-api/0.log" Dec 04 05:15:12 crc kubenswrapper[4806]: I1204 05:15:12.158994 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-949b6795d-z7dx2_ef61aa0b-2b2c-481f-b22d-4ea770a91711/barbican-api-log/0.log" Dec 04 05:15:12 crc kubenswrapper[4806]: I1204 05:15:12.188117 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-b658d66b-ft4r8_fef48eb8-ea9d-42ab-8839-fa7a890fe301/barbican-keystone-listener/0.log" Dec 04 05:15:12 crc kubenswrapper[4806]: I1204 05:15:12.284746 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-b658d66b-ft4r8_fef48eb8-ea9d-42ab-8839-fa7a890fe301/barbican-keystone-listener-log/0.log" Dec 04 05:15:12 crc kubenswrapper[4806]: I1204 05:15:12.400300 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6785cc669c-hgz8x_cc467d3b-daf0-4869-9cc3-1034cdfd373a/barbican-worker-log/0.log" Dec 04 05:15:12 crc kubenswrapper[4806]: I1204 05:15:12.434644 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6785cc669c-hgz8x_cc467d3b-daf0-4869-9cc3-1034cdfd373a/barbican-worker/0.log" Dec 04 05:15:12 crc kubenswrapper[4806]: I1204 05:15:12.933133 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8_1787b37a-3067-4912-a658-4a8fd53faab6/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:15:12 crc kubenswrapper[4806]: I1204 05:15:12.982871 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_56b1c86b-9508-4ca9-a2b8-9a496875c9e8/ceilometer-central-agent/0.log" Dec 04 05:15:13 crc kubenswrapper[4806]: I1204 05:15:13.127687 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_56b1c86b-9508-4ca9-a2b8-9a496875c9e8/proxy-httpd/0.log" Dec 04 05:15:13 crc kubenswrapper[4806]: I1204 05:15:13.164415 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_56b1c86b-9508-4ca9-a2b8-9a496875c9e8/sg-core/0.log" Dec 04 05:15:13 crc kubenswrapper[4806]: I1204 05:15:13.187813 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_56b1c86b-9508-4ca9-a2b8-9a496875c9e8/ceilometer-notification-agent/0.log" Dec 04 05:15:13 crc kubenswrapper[4806]: I1204 05:15:13.416855 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3d05a5b8-2294-4d30-a0b7-d865c8aae646/cinder-api/0.log" Dec 04 05:15:13 crc kubenswrapper[4806]: I1204 05:15:13.451878 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3d05a5b8-2294-4d30-a0b7-d865c8aae646/cinder-api-log/0.log" Dec 04 05:15:13 crc kubenswrapper[4806]: I1204 05:15:13.617245 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8205f150-c0db-445a-b005-30d0632610c0/cinder-scheduler/0.log" Dec 04 05:15:13 crc kubenswrapper[4806]: I1204 05:15:13.697279 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8205f150-c0db-445a-b005-30d0632610c0/probe/0.log" Dec 04 05:15:13 crc kubenswrapper[4806]: I1204 05:15:13.804009 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6_bbebb80f-3f15-464c-bb52-ab7dc2a100c4/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:15:13 crc kubenswrapper[4806]: I1204 05:15:13.966740 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-984f8_cc3788a8-d331-4c5c-a80b-d73e1ef5da9c/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:15:14 crc kubenswrapper[4806]: I1204 05:15:14.036145 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7677974f-lk2tx_5e42fed2-6410-480b-ad51-7c8a5c383eea/init/0.log" Dec 04 05:15:14 crc kubenswrapper[4806]: I1204 05:15:14.330695 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7677974f-lk2tx_5e42fed2-6410-480b-ad51-7c8a5c383eea/init/0.log" Dec 04 05:15:14 crc kubenswrapper[4806]: I1204 05:15:14.427685 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv_0b8e6232-8e76-425c-9b02-94e0a3c236cb/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:15:14 crc kubenswrapper[4806]: I1204 05:15:14.526453 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7677974f-lk2tx_5e42fed2-6410-480b-ad51-7c8a5c383eea/dnsmasq-dns/0.log" Dec 04 05:15:14 crc kubenswrapper[4806]: I1204 05:15:14.802754 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0f1138ab-7d74-4dc5-883c-e32dd1f21546/glance-log/0.log" Dec 04 05:15:14 crc kubenswrapper[4806]: I1204 05:15:14.813697 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0f1138ab-7d74-4dc5-883c-e32dd1f21546/glance-httpd/0.log" Dec 04 05:15:15 crc kubenswrapper[4806]: I1204 05:15:15.051569 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_5d515cfb-b26c-430f-9ee9-68b7d1a87fb6/glance-httpd/0.log" Dec 04 05:15:15 crc kubenswrapper[4806]: I1204 05:15:15.069052 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_5d515cfb-b26c-430f-9ee9-68b7d1a87fb6/glance-log/0.log" Dec 04 05:15:15 crc kubenswrapper[4806]: I1204 05:15:15.373383 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d8bd6c96d-4sqhf_d2d53864-c3c8-4e40-8791-0bac49ba561e/horizon/2.log" Dec 04 05:15:15 crc kubenswrapper[4806]: I1204 05:15:15.374746 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d8bd6c96d-4sqhf_d2d53864-c3c8-4e40-8791-0bac49ba561e/horizon/1.log" Dec 04 05:15:15 crc kubenswrapper[4806]: I1204 05:15:15.700547 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j_f3284126-2542-43fc-8e96-829f03db14bc/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:15:15 crc kubenswrapper[4806]: I1204 05:15:15.813123 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-fdr25_3a53f8ff-b197-4f2b-8194-6424a21dae22/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:15:15 crc kubenswrapper[4806]: I1204 05:15:15.929235 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d8bd6c96d-4sqhf_d2d53864-c3c8-4e40-8791-0bac49ba561e/horizon-log/0.log" Dec 04 05:15:16 crc kubenswrapper[4806]: I1204 05:15:16.245453 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29413741-vfktv_106a7dfc-57ef-4177-821d-f77a0329ac90/keystone-cron/0.log" Dec 04 05:15:16 crc kubenswrapper[4806]: I1204 05:15:16.448191 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_00dc1c6d-f6cc-49ac-935f-24007536925c/kube-state-metrics/0.log" Dec 04 05:15:16 crc kubenswrapper[4806]: I1204 05:15:16.569254 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-64cc7bf6f4-fcxhm_0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5/keystone-api/0.log" Dec 04 05:15:16 crc kubenswrapper[4806]: I1204 05:15:16.733619 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2_daec7717-fd5a-47be-a766-1ef8bbabab14/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:15:17 crc kubenswrapper[4806]: I1204 05:15:17.400230 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq_4b974095-c38a-43c6-a6ac-67497d4f509f/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:15:17 crc kubenswrapper[4806]: I1204 05:15:17.512024 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-669cc9c8df-p88wk_7fbb687c-667a-4025-bd3c-1eb037758c0d/neutron-httpd/0.log" Dec 04 05:15:17 crc kubenswrapper[4806]: I1204 05:15:17.615791 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-669cc9c8df-p88wk_7fbb687c-667a-4025-bd3c-1eb037758c0d/neutron-api/0.log" Dec 04 05:15:18 crc kubenswrapper[4806]: I1204 05:15:18.268067 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_713e59da-b517-4099-bf3e-5be89d93e82c/nova-cell0-conductor-conductor/0.log" Dec 04 05:15:18 crc kubenswrapper[4806]: I1204 05:15:18.592581 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_e7137b5c-7316-46a0-abda-976ee89ac4ac/nova-cell1-conductor-conductor/0.log" Dec 04 05:15:18 crc kubenswrapper[4806]: I1204 05:15:18.936084 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_3999d3e5-7b66-4792-9fd0-996d0d97efd2/nova-cell1-novncproxy-novncproxy/0.log" Dec 04 05:15:19 crc kubenswrapper[4806]: I1204 05:15:19.153996 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4d149e6e-bdab-4137-a48a-27ed21680ab8/nova-api-log/0.log" Dec 04 05:15:19 crc kubenswrapper[4806]: I1204 05:15:19.274649 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-l4hb8_21f3ccbb-669c-4055-b85c-f0c69fe61a8e/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:15:19 crc kubenswrapper[4806]: I1204 05:15:19.526253 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac/nova-metadata-log/0.log" Dec 04 05:15:19 crc kubenswrapper[4806]: I1204 05:15:19.549037 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4d149e6e-bdab-4137-a48a-27ed21680ab8/nova-api-api/0.log" Dec 04 05:15:19 crc kubenswrapper[4806]: I1204 05:15:19.974203 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0d1a2b83-04d7-4db8-acd1-456cc0c33206/mysql-bootstrap/0.log" Dec 04 05:15:20 crc kubenswrapper[4806]: I1204 05:15:20.175038 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0d1a2b83-04d7-4db8-acd1-456cc0c33206/mysql-bootstrap/0.log" Dec 04 05:15:20 crc kubenswrapper[4806]: I1204 05:15:20.223635 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0d1a2b83-04d7-4db8-acd1-456cc0c33206/galera/0.log" Dec 04 05:15:20 crc kubenswrapper[4806]: I1204 05:15:20.349712 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_a32a7b4e-8881-437b-843e-54a6d1e5c11e/nova-scheduler-scheduler/0.log" Dec 04 05:15:20 crc kubenswrapper[4806]: I1204 05:15:20.616979 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_99326fd5-69d5-4e7d-b5c6-f7808344ad6d/mysql-bootstrap/0.log" Dec 04 05:15:20 crc kubenswrapper[4806]: I1204 05:15:20.706542 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_99326fd5-69d5-4e7d-b5c6-f7808344ad6d/galera/0.log" Dec 04 05:15:20 crc kubenswrapper[4806]: I1204 05:15:20.812052 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_99326fd5-69d5-4e7d-b5c6-f7808344ad6d/mysql-bootstrap/0.log" Dec 04 05:15:20 crc kubenswrapper[4806]: I1204 05:15:20.930722 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_266dbc23-c65d-4b9d-978d-09f9de4a06e5/openstackclient/0.log" Dec 04 05:15:21 crc kubenswrapper[4806]: I1204 05:15:21.221609 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-4kl52_0b91c83f-0036-42a8-97b4-86846f14466a/openstack-network-exporter/0.log" Dec 04 05:15:21 crc kubenswrapper[4806]: I1204 05:15:21.326644 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac/nova-metadata-metadata/0.log" Dec 04 05:15:21 crc kubenswrapper[4806]: I1204 05:15:21.414705 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2sz9t_63e4ca21-d6b3-4e93-965e-3dde5e35027f/ovsdb-server-init/0.log" Dec 04 05:15:21 crc kubenswrapper[4806]: I1204 05:15:21.606642 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2sz9t_63e4ca21-d6b3-4e93-965e-3dde5e35027f/ovs-vswitchd/0.log" Dec 04 05:15:21 crc kubenswrapper[4806]: I1204 05:15:21.607947 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2sz9t_63e4ca21-d6b3-4e93-965e-3dde5e35027f/ovsdb-server-init/0.log" Dec 04 05:15:21 crc kubenswrapper[4806]: I1204 05:15:21.693395 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2sz9t_63e4ca21-d6b3-4e93-965e-3dde5e35027f/ovsdb-server/0.log" Dec 04 05:15:21 crc kubenswrapper[4806]: I1204 05:15:21.959066 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-tx487_ef210587-097b-48d8-8097-dc1b95e2456e/ovn-controller/0.log" Dec 04 05:15:22 crc kubenswrapper[4806]: I1204 05:15:22.026533 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-mbhj2_b4d109a9-916a-41b6-9008-15291d44a0a4/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:15:22 crc kubenswrapper[4806]: I1204 05:15:22.214053 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_55f867c0-7de7-40bc-839a-4f9d9d09dd2e/ovn-northd/0.log" Dec 04 05:15:22 crc kubenswrapper[4806]: I1204 05:15:22.217664 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_55f867c0-7de7-40bc-839a-4f9d9d09dd2e/openstack-network-exporter/0.log" Dec 04 05:15:22 crc kubenswrapper[4806]: I1204 05:15:22.410794 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_75e0b5e3-e994-4ef3-9673-539ca96485c5/openstack-network-exporter/0.log" Dec 04 05:15:22 crc kubenswrapper[4806]: I1204 05:15:22.454780 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_75e0b5e3-e994-4ef3-9673-539ca96485c5/ovsdbserver-nb/0.log" Dec 04 05:15:22 crc kubenswrapper[4806]: I1204 05:15:22.631523 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_09739947-6f7d-4a4e-8a85-05a8a7345894/openstack-network-exporter/0.log" Dec 04 05:15:22 crc kubenswrapper[4806]: I1204 05:15:22.682493 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_09739947-6f7d-4a4e-8a85-05a8a7345894/ovsdbserver-sb/0.log" Dec 04 05:15:23 crc kubenswrapper[4806]: I1204 05:15:23.044446 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_17d9cfc5-f7f9-442f-8174-2b61041b8f70/setup-container/0.log" Dec 04 05:15:23 crc kubenswrapper[4806]: I1204 05:15:23.155010 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-d9c876cd8-zfctt_122ae99a-2891-4bdb-903a-9cccfb3df2a1/placement-api/0.log" Dec 04 05:15:23 crc kubenswrapper[4806]: I1204 05:15:23.229324 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-d9c876cd8-zfctt_122ae99a-2891-4bdb-903a-9cccfb3df2a1/placement-log/0.log" Dec 04 05:15:23 crc kubenswrapper[4806]: I1204 05:15:23.429633 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_17d9cfc5-f7f9-442f-8174-2b61041b8f70/rabbitmq/0.log" Dec 04 05:15:23 crc kubenswrapper[4806]: I1204 05:15:23.436476 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_17d9cfc5-f7f9-442f-8174-2b61041b8f70/setup-container/0.log" Dec 04 05:15:23 crc kubenswrapper[4806]: I1204 05:15:23.553248 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5a2a4a0f-f385-4fad-9825-2824f47dd8db/setup-container/0.log" Dec 04 05:15:23 crc kubenswrapper[4806]: I1204 05:15:23.772070 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5a2a4a0f-f385-4fad-9825-2824f47dd8db/rabbitmq/0.log" Dec 04 05:15:23 crc kubenswrapper[4806]: I1204 05:15:23.785100 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5a2a4a0f-f385-4fad-9825-2824f47dd8db/setup-container/0.log" Dec 04 05:15:23 crc kubenswrapper[4806]: I1204 05:15:23.852663 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-fprln_a74530eb-f7f4-40f7-b1fc-c66cba904ce5/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:15:24 crc kubenswrapper[4806]: I1204 05:15:24.129747 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld_410d8040-e0b5-4869-bb16-41e5b59487d3/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:15:24 crc kubenswrapper[4806]: I1204 05:15:24.192319 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-ds7xn_ef28af41-0fb7-4e9a-adbe-c08fdf320e95/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:15:24 crc kubenswrapper[4806]: I1204 05:15:24.394470 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-25667_e5f894f3-1fed-46cc-9caa-8f61712e8ab3/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:15:24 crc kubenswrapper[4806]: I1204 05:15:24.493775 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-cddcl_cf2d196e-c288-449d-a106-2ca0b6697582/ssh-known-hosts-edpm-deployment/0.log" Dec 04 05:15:25 crc kubenswrapper[4806]: I1204 05:15:25.094354 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-685c486f6f-6zmmj_794d2127-4482-4196-b558-d24f003e96ca/proxy-server/0.log" Dec 04 05:15:25 crc kubenswrapper[4806]: I1204 05:15:25.299152 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-685c486f6f-6zmmj_794d2127-4482-4196-b558-d24f003e96ca/proxy-httpd/0.log" Dec 04 05:15:25 crc kubenswrapper[4806]: I1204 05:15:25.348222 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-t7k9p_85f9dc59-e013-464d-b305-77c85cb82f4a/swift-ring-rebalance/0.log" Dec 04 05:15:25 crc kubenswrapper[4806]: I1204 05:15:25.462937 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/account-auditor/0.log" Dec 04 05:15:25 crc kubenswrapper[4806]: I1204 05:15:25.620014 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/account-replicator/0.log" Dec 04 05:15:25 crc kubenswrapper[4806]: I1204 05:15:25.636170 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/account-reaper/0.log" Dec 04 05:15:25 crc kubenswrapper[4806]: I1204 05:15:25.661317 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/account-server/0.log" Dec 04 05:15:25 crc kubenswrapper[4806]: I1204 05:15:25.675781 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/container-auditor/0.log" Dec 04 05:15:25 crc kubenswrapper[4806]: I1204 05:15:25.852853 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/container-server/0.log" Dec 04 05:15:25 crc kubenswrapper[4806]: I1204 05:15:25.873419 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/container-replicator/0.log" Dec 04 05:15:25 crc kubenswrapper[4806]: I1204 05:15:25.937262 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/container-updater/0.log" Dec 04 05:15:26 crc kubenswrapper[4806]: I1204 05:15:26.003125 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/object-auditor/0.log" Dec 04 05:15:26 crc kubenswrapper[4806]: I1204 05:15:26.634459 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/object-expirer/0.log" Dec 04 05:15:26 crc kubenswrapper[4806]: I1204 05:15:26.743358 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/object-updater/0.log" Dec 04 05:15:26 crc kubenswrapper[4806]: I1204 05:15:26.759277 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/object-replicator/0.log" Dec 04 05:15:26 crc kubenswrapper[4806]: I1204 05:15:26.817846 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/object-server/0.log" Dec 04 05:15:26 crc kubenswrapper[4806]: I1204 05:15:26.950504 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/rsync/0.log" Dec 04 05:15:27 crc kubenswrapper[4806]: I1204 05:15:27.038425 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/swift-recon-cron/0.log" Dec 04 05:15:27 crc kubenswrapper[4806]: I1204 05:15:27.343673 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-stpm5_c9247c4e-1c6a-4110-8d97-6829c6064d48/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:15:27 crc kubenswrapper[4806]: I1204 05:15:27.473297 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_319a7357-70c6-4b54-8e97-9860dd0618e7/tempest-tests-tempest-tests-runner/0.log" Dec 04 05:15:27 crc kubenswrapper[4806]: I1204 05:15:27.670756 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_4922f2b6-05e3-4f78-a14a-46ca2668edfb/test-operator-logs-container/0.log" Dec 04 05:15:27 crc kubenswrapper[4806]: I1204 05:15:27.828854 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-phq5h_02021c61-fa0c-4aea-80d4-b0eb592c0d06/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:15:38 crc kubenswrapper[4806]: I1204 05:15:38.401721 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_0face013-e59b-4e32-ace5-d64d45302e7c/memcached/0.log" Dec 04 05:15:56 crc kubenswrapper[4806]: I1204 05:15:56.655686 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh_a72e0bcd-e604-4243-b294-70ebe7ba990c/util/0.log" Dec 04 05:15:56 crc kubenswrapper[4806]: I1204 05:15:56.828484 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh_a72e0bcd-e604-4243-b294-70ebe7ba990c/util/0.log" Dec 04 05:15:56 crc kubenswrapper[4806]: I1204 05:15:56.936607 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh_a72e0bcd-e604-4243-b294-70ebe7ba990c/pull/0.log" Dec 04 05:15:56 crc kubenswrapper[4806]: I1204 05:15:56.947049 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh_a72e0bcd-e604-4243-b294-70ebe7ba990c/pull/0.log" Dec 04 05:15:57 crc kubenswrapper[4806]: I1204 05:15:57.047360 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:15:57 crc kubenswrapper[4806]: I1204 05:15:57.047411 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:15:57 crc kubenswrapper[4806]: I1204 05:15:57.084497 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh_a72e0bcd-e604-4243-b294-70ebe7ba990c/util/0.log" Dec 04 05:15:57 crc kubenswrapper[4806]: I1204 05:15:57.124217 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh_a72e0bcd-e604-4243-b294-70ebe7ba990c/pull/0.log" Dec 04 05:15:57 crc kubenswrapper[4806]: I1204 05:15:57.139902 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh_a72e0bcd-e604-4243-b294-70ebe7ba990c/extract/0.log" Dec 04 05:15:57 crc kubenswrapper[4806]: I1204 05:15:57.320979 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-g2dfw_b2bba37e-d570-432f-9520-9b5de2005e5c/kube-rbac-proxy/0.log" Dec 04 05:15:57 crc kubenswrapper[4806]: I1204 05:15:57.390637 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-g2dfw_b2bba37e-d570-432f-9520-9b5de2005e5c/manager/0.log" Dec 04 05:15:57 crc kubenswrapper[4806]: I1204 05:15:57.429059 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-25nxx_a022d699-4a2b-49e2-8d8f-787a3d888df8/kube-rbac-proxy/0.log" Dec 04 05:15:57 crc kubenswrapper[4806]: I1204 05:15:57.562868 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-25nxx_a022d699-4a2b-49e2-8d8f-787a3d888df8/manager/0.log" Dec 04 05:15:57 crc kubenswrapper[4806]: I1204 05:15:57.635203 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-mxmf2_91ab6e01-027e-4e7a-a250-374b4e8873fc/kube-rbac-proxy/0.log" Dec 04 05:15:57 crc kubenswrapper[4806]: I1204 05:15:57.659895 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-mxmf2_91ab6e01-027e-4e7a-a250-374b4e8873fc/manager/0.log" Dec 04 05:15:57 crc kubenswrapper[4806]: I1204 05:15:57.840121 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-xd6pj_7213e0bc-21af-4df7-929d-f915e343f184/kube-rbac-proxy/0.log" Dec 04 05:15:57 crc kubenswrapper[4806]: I1204 05:15:57.897896 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-xd6pj_7213e0bc-21af-4df7-929d-f915e343f184/manager/0.log" Dec 04 05:15:58 crc kubenswrapper[4806]: I1204 05:15:58.063539 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-vhxjt_b496f26d-2493-416f-a0f8-5b33e41dbda0/kube-rbac-proxy/0.log" Dec 04 05:15:58 crc kubenswrapper[4806]: I1204 05:15:58.089170 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-vhxjt_b496f26d-2493-416f-a0f8-5b33e41dbda0/manager/0.log" Dec 04 05:15:58 crc kubenswrapper[4806]: I1204 05:15:58.690537 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-vtcg5_084c526b-94d8-43ae-9ffb-33aa215de8c2/manager/0.log" Dec 04 05:15:58 crc kubenswrapper[4806]: I1204 05:15:58.734520 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-g6svq_04d2c949-c17a-4299-a9be-b56a7c34866c/kube-rbac-proxy/0.log" Dec 04 05:15:58 crc kubenswrapper[4806]: I1204 05:15:58.760999 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-vtcg5_084c526b-94d8-43ae-9ffb-33aa215de8c2/kube-rbac-proxy/0.log" Dec 04 05:15:59 crc kubenswrapper[4806]: I1204 05:15:59.032400 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-f6k5f_5f55ddcf-ca5d-4249-9384-e541ae7d4536/manager/0.log" Dec 04 05:15:59 crc kubenswrapper[4806]: I1204 05:15:59.037363 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-f6k5f_5f55ddcf-ca5d-4249-9384-e541ae7d4536/kube-rbac-proxy/0.log" Dec 04 05:15:59 crc kubenswrapper[4806]: I1204 05:15:59.043027 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-g6svq_04d2c949-c17a-4299-a9be-b56a7c34866c/manager/0.log" Dec 04 05:15:59 crc kubenswrapper[4806]: I1204 05:15:59.271224 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-nt5vq_669dd56a-69b0-4315-aacc-c8f085f0b76a/kube-rbac-proxy/0.log" Dec 04 05:15:59 crc kubenswrapper[4806]: I1204 05:15:59.351774 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-nt5vq_669dd56a-69b0-4315-aacc-c8f085f0b76a/manager/0.log" Dec 04 05:15:59 crc kubenswrapper[4806]: I1204 05:15:59.500817 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-2spbg_66428f64-e34e-4b37-b911-13e7142f1c18/kube-rbac-proxy/0.log" Dec 04 05:15:59 crc kubenswrapper[4806]: I1204 05:15:59.590595 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-2spbg_66428f64-e34e-4b37-b911-13e7142f1c18/manager/0.log" Dec 04 05:16:00 crc kubenswrapper[4806]: I1204 05:16:00.323878 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-45jzr_310d58eb-6345-4cba-8a01-373c65c7bc40/manager/0.log" Dec 04 05:16:00 crc kubenswrapper[4806]: I1204 05:16:00.331505 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-45jzr_310d58eb-6345-4cba-8a01-373c65c7bc40/kube-rbac-proxy/0.log" Dec 04 05:16:00 crc kubenswrapper[4806]: I1204 05:16:00.553281 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-2b2qk_95ae1fa1-3ac7-4040-86da-150daca97f6b/manager/0.log" Dec 04 05:16:00 crc kubenswrapper[4806]: I1204 05:16:00.621003 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-pk7ws_5f479a93-bb3a-45b8-aa74-60e9f0a7aebd/kube-rbac-proxy/0.log" Dec 04 05:16:00 crc kubenswrapper[4806]: I1204 05:16:00.621660 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-2b2qk_95ae1fa1-3ac7-4040-86da-150daca97f6b/kube-rbac-proxy/0.log" Dec 04 05:16:00 crc kubenswrapper[4806]: I1204 05:16:00.792582 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-pk7ws_5f479a93-bb3a-45b8-aa74-60e9f0a7aebd/manager/0.log" Dec 04 05:16:00 crc kubenswrapper[4806]: I1204 05:16:00.851517 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-kcmvz_1cfe7f94-e830-4145-a3e7-8a3f694adda1/kube-rbac-proxy/0.log" Dec 04 05:16:00 crc kubenswrapper[4806]: I1204 05:16:00.862874 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-kcmvz_1cfe7f94-e830-4145-a3e7-8a3f694adda1/manager/0.log" Dec 04 05:16:01 crc kubenswrapper[4806]: I1204 05:16:01.039811 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7_2a7907e6-c36f-4534-9460-80c0091af286/kube-rbac-proxy/0.log" Dec 04 05:16:01 crc kubenswrapper[4806]: I1204 05:16:01.041371 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7_2a7907e6-c36f-4534-9460-80c0091af286/manager/0.log" Dec 04 05:16:01 crc kubenswrapper[4806]: I1204 05:16:01.451398 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7bfcb77fcc-bx765_4d229897-71a8-4b61-a9be-4d963e1384ac/operator/0.log" Dec 04 05:16:01 crc kubenswrapper[4806]: I1204 05:16:01.583346 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-gsvwx_89d5c50c-5d00-4f21-9e60-8c6692a0ad33/registry-server/0.log" Dec 04 05:16:01 crc kubenswrapper[4806]: I1204 05:16:01.794681 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-sktq6_e06aa87c-9f7d-44f7-b96d-27141d2fff93/kube-rbac-proxy/0.log" Dec 04 05:16:01 crc kubenswrapper[4806]: I1204 05:16:01.947240 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-mnhlt_45192fbb-caf2-4df8-82a6-c222f73b033c/kube-rbac-proxy/0.log" Dec 04 05:16:02 crc kubenswrapper[4806]: I1204 05:16:02.017401 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-sktq6_e06aa87c-9f7d-44f7-b96d-27141d2fff93/manager/0.log" Dec 04 05:16:02 crc kubenswrapper[4806]: I1204 05:16:02.186692 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-mnhlt_45192fbb-caf2-4df8-82a6-c222f73b033c/manager/0.log" Dec 04 05:16:02 crc kubenswrapper[4806]: I1204 05:16:02.284188 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8d698776b-npcz2_ff1a6781-908b-419f-8a02-10abbbfa0816/manager/0.log" Dec 04 05:16:02 crc kubenswrapper[4806]: I1204 05:16:02.297388 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-cwnth_be179b75-6e57-4537-b3ee-a57dc049948e/operator/0.log" Dec 04 05:16:02 crc kubenswrapper[4806]: I1204 05:16:02.400664 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-fdqbx_3580e441-1df2-4fe9-8969-e1621fc554dc/kube-rbac-proxy/0.log" Dec 04 05:16:02 crc kubenswrapper[4806]: I1204 05:16:02.468064 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-fdqbx_3580e441-1df2-4fe9-8969-e1621fc554dc/manager/0.log" Dec 04 05:16:02 crc kubenswrapper[4806]: I1204 05:16:02.526295 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-kvnjv_647621de-6ecf-4ebe-96f2-e85f26e4917e/kube-rbac-proxy/0.log" Dec 04 05:16:02 crc kubenswrapper[4806]: I1204 05:16:02.568173 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-kvnjv_647621de-6ecf-4ebe-96f2-e85f26e4917e/manager/0.log" Dec 04 05:16:02 crc kubenswrapper[4806]: I1204 05:16:02.674722 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-69l55_2473e4f1-90f0-43a0-9e72-0c761d335d76/kube-rbac-proxy/0.log" Dec 04 05:16:02 crc kubenswrapper[4806]: I1204 05:16:02.791040 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-2lk62_a3827706-d53c-48f3-b4cf-1c240c0f868c/kube-rbac-proxy/0.log" Dec 04 05:16:02 crc kubenswrapper[4806]: I1204 05:16:02.796488 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-69l55_2473e4f1-90f0-43a0-9e72-0c761d335d76/manager/0.log" Dec 04 05:16:02 crc kubenswrapper[4806]: I1204 05:16:02.871289 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-2lk62_a3827706-d53c-48f3-b4cf-1c240c0f868c/manager/0.log" Dec 04 05:16:03 crc kubenswrapper[4806]: I1204 05:16:03.144526 4806 scope.go:117] "RemoveContainer" containerID="383c717009bb281750543842f95b30ff9a9ead8910583ddc5ad438fc5e444db1" Dec 04 05:16:19 crc kubenswrapper[4806]: I1204 05:16:19.090902 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4vvqv"] Dec 04 05:16:19 crc kubenswrapper[4806]: E1204 05:16:19.091817 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c37fbb2-7a56-413d-a7a3-67c6adc26608" containerName="collect-profiles" Dec 04 05:16:19 crc kubenswrapper[4806]: I1204 05:16:19.091829 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c37fbb2-7a56-413d-a7a3-67c6adc26608" containerName="collect-profiles" Dec 04 05:16:19 crc kubenswrapper[4806]: I1204 05:16:19.093908 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c37fbb2-7a56-413d-a7a3-67c6adc26608" containerName="collect-profiles" Dec 04 05:16:19 crc kubenswrapper[4806]: I1204 05:16:19.095369 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:19 crc kubenswrapper[4806]: I1204 05:16:19.105917 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4vvqv"] Dec 04 05:16:19 crc kubenswrapper[4806]: I1204 05:16:19.177743 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/206f5ad4-5290-4ac9-86c2-b4e945849790-utilities\") pod \"redhat-marketplace-4vvqv\" (UID: \"206f5ad4-5290-4ac9-86c2-b4e945849790\") " pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:19 crc kubenswrapper[4806]: I1204 05:16:19.177785 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/206f5ad4-5290-4ac9-86c2-b4e945849790-catalog-content\") pod \"redhat-marketplace-4vvqv\" (UID: \"206f5ad4-5290-4ac9-86c2-b4e945849790\") " pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:19 crc kubenswrapper[4806]: I1204 05:16:19.178243 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4nzm\" (UniqueName: \"kubernetes.io/projected/206f5ad4-5290-4ac9-86c2-b4e945849790-kube-api-access-g4nzm\") pod \"redhat-marketplace-4vvqv\" (UID: \"206f5ad4-5290-4ac9-86c2-b4e945849790\") " pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:19 crc kubenswrapper[4806]: I1204 05:16:19.279761 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4nzm\" (UniqueName: \"kubernetes.io/projected/206f5ad4-5290-4ac9-86c2-b4e945849790-kube-api-access-g4nzm\") pod \"redhat-marketplace-4vvqv\" (UID: \"206f5ad4-5290-4ac9-86c2-b4e945849790\") " pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:19 crc kubenswrapper[4806]: I1204 05:16:19.279850 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/206f5ad4-5290-4ac9-86c2-b4e945849790-utilities\") pod \"redhat-marketplace-4vvqv\" (UID: \"206f5ad4-5290-4ac9-86c2-b4e945849790\") " pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:19 crc kubenswrapper[4806]: I1204 05:16:19.279866 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/206f5ad4-5290-4ac9-86c2-b4e945849790-catalog-content\") pod \"redhat-marketplace-4vvqv\" (UID: \"206f5ad4-5290-4ac9-86c2-b4e945849790\") " pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:19 crc kubenswrapper[4806]: I1204 05:16:19.280407 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/206f5ad4-5290-4ac9-86c2-b4e945849790-catalog-content\") pod \"redhat-marketplace-4vvqv\" (UID: \"206f5ad4-5290-4ac9-86c2-b4e945849790\") " pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:19 crc kubenswrapper[4806]: I1204 05:16:19.280879 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/206f5ad4-5290-4ac9-86c2-b4e945849790-utilities\") pod \"redhat-marketplace-4vvqv\" (UID: \"206f5ad4-5290-4ac9-86c2-b4e945849790\") " pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:19 crc kubenswrapper[4806]: I1204 05:16:19.301193 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4nzm\" (UniqueName: \"kubernetes.io/projected/206f5ad4-5290-4ac9-86c2-b4e945849790-kube-api-access-g4nzm\") pod \"redhat-marketplace-4vvqv\" (UID: \"206f5ad4-5290-4ac9-86c2-b4e945849790\") " pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:19 crc kubenswrapper[4806]: I1204 05:16:19.446524 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:19 crc kubenswrapper[4806]: I1204 05:16:19.951775 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4vvqv"] Dec 04 05:16:20 crc kubenswrapper[4806]: I1204 05:16:20.222808 4806 generic.go:334] "Generic (PLEG): container finished" podID="206f5ad4-5290-4ac9-86c2-b4e945849790" containerID="05bca03557fd83437b470a8caec95e2382ecbb2642c1ba4fba9dbb261bd05a1e" exitCode=0 Dec 04 05:16:20 crc kubenswrapper[4806]: I1204 05:16:20.222849 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vvqv" event={"ID":"206f5ad4-5290-4ac9-86c2-b4e945849790","Type":"ContainerDied","Data":"05bca03557fd83437b470a8caec95e2382ecbb2642c1ba4fba9dbb261bd05a1e"} Dec 04 05:16:20 crc kubenswrapper[4806]: I1204 05:16:20.222878 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vvqv" event={"ID":"206f5ad4-5290-4ac9-86c2-b4e945849790","Type":"ContainerStarted","Data":"d5a2befc60b57b40209ddf179cbae34ebecea2ec37229cb789d5631a5dce1ef0"} Dec 04 05:16:21 crc kubenswrapper[4806]: I1204 05:16:21.235133 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vvqv" event={"ID":"206f5ad4-5290-4ac9-86c2-b4e945849790","Type":"ContainerStarted","Data":"9a2cfd286a0070df72755905336e44e402924e05f777ff733f967a10c8bfb2a5"} Dec 04 05:16:22 crc kubenswrapper[4806]: I1204 05:16:22.243638 4806 generic.go:334] "Generic (PLEG): container finished" podID="206f5ad4-5290-4ac9-86c2-b4e945849790" containerID="9a2cfd286a0070df72755905336e44e402924e05f777ff733f967a10c8bfb2a5" exitCode=0 Dec 04 05:16:22 crc kubenswrapper[4806]: I1204 05:16:22.243690 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vvqv" event={"ID":"206f5ad4-5290-4ac9-86c2-b4e945849790","Type":"ContainerDied","Data":"9a2cfd286a0070df72755905336e44e402924e05f777ff733f967a10c8bfb2a5"} Dec 04 05:16:23 crc kubenswrapper[4806]: I1204 05:16:23.256233 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vvqv" event={"ID":"206f5ad4-5290-4ac9-86c2-b4e945849790","Type":"ContainerStarted","Data":"3e080fc639e263395b2252ec5afdcb24a3e6717ffe0af47960f0602921ec3519"} Dec 04 05:16:24 crc kubenswrapper[4806]: I1204 05:16:24.468491 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-dgkrj_0647d456-31a3-4906-9d62-f43b4cfef5cb/control-plane-machine-set-operator/0.log" Dec 04 05:16:24 crc kubenswrapper[4806]: I1204 05:16:24.699014 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j8m72_7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6/kube-rbac-proxy/0.log" Dec 04 05:16:24 crc kubenswrapper[4806]: I1204 05:16:24.771229 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j8m72_7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6/machine-api-operator/0.log" Dec 04 05:16:27 crc kubenswrapper[4806]: I1204 05:16:27.047215 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:16:27 crc kubenswrapper[4806]: I1204 05:16:27.047579 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:16:29 crc kubenswrapper[4806]: I1204 05:16:29.447888 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:29 crc kubenswrapper[4806]: I1204 05:16:29.448196 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:29 crc kubenswrapper[4806]: I1204 05:16:29.553995 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:29 crc kubenswrapper[4806]: I1204 05:16:29.574400 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4vvqv" podStartSLOduration=8.172794884 podStartE2EDuration="10.574375498s" podCreationTimestamp="2025-12-04 05:16:19 +0000 UTC" firstStartedPulling="2025-12-04 05:16:20.2248593 +0000 UTC m=+4895.083372228" lastFinishedPulling="2025-12-04 05:16:22.626439914 +0000 UTC m=+4897.484952842" observedRunningTime="2025-12-04 05:16:23.282618998 +0000 UTC m=+4898.141131926" watchObservedRunningTime="2025-12-04 05:16:29.574375498 +0000 UTC m=+4904.432888426" Dec 04 05:16:30 crc kubenswrapper[4806]: I1204 05:16:30.357345 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:30 crc kubenswrapper[4806]: I1204 05:16:30.406337 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4vvqv"] Dec 04 05:16:32 crc kubenswrapper[4806]: I1204 05:16:32.328628 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4vvqv" podUID="206f5ad4-5290-4ac9-86c2-b4e945849790" containerName="registry-server" containerID="cri-o://3e080fc639e263395b2252ec5afdcb24a3e6717ffe0af47960f0602921ec3519" gracePeriod=2 Dec 04 05:16:32 crc kubenswrapper[4806]: I1204 05:16:32.785403 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:32 crc kubenswrapper[4806]: I1204 05:16:32.885734 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/206f5ad4-5290-4ac9-86c2-b4e945849790-utilities\") pod \"206f5ad4-5290-4ac9-86c2-b4e945849790\" (UID: \"206f5ad4-5290-4ac9-86c2-b4e945849790\") " Dec 04 05:16:32 crc kubenswrapper[4806]: I1204 05:16:32.885899 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4nzm\" (UniqueName: \"kubernetes.io/projected/206f5ad4-5290-4ac9-86c2-b4e945849790-kube-api-access-g4nzm\") pod \"206f5ad4-5290-4ac9-86c2-b4e945849790\" (UID: \"206f5ad4-5290-4ac9-86c2-b4e945849790\") " Dec 04 05:16:32 crc kubenswrapper[4806]: I1204 05:16:32.886059 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/206f5ad4-5290-4ac9-86c2-b4e945849790-catalog-content\") pod \"206f5ad4-5290-4ac9-86c2-b4e945849790\" (UID: \"206f5ad4-5290-4ac9-86c2-b4e945849790\") " Dec 04 05:16:32 crc kubenswrapper[4806]: I1204 05:16:32.886481 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/206f5ad4-5290-4ac9-86c2-b4e945849790-utilities" (OuterVolumeSpecName: "utilities") pod "206f5ad4-5290-4ac9-86c2-b4e945849790" (UID: "206f5ad4-5290-4ac9-86c2-b4e945849790"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:16:32 crc kubenswrapper[4806]: I1204 05:16:32.906384 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/206f5ad4-5290-4ac9-86c2-b4e945849790-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "206f5ad4-5290-4ac9-86c2-b4e945849790" (UID: "206f5ad4-5290-4ac9-86c2-b4e945849790"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:16:32 crc kubenswrapper[4806]: I1204 05:16:32.911293 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/206f5ad4-5290-4ac9-86c2-b4e945849790-kube-api-access-g4nzm" (OuterVolumeSpecName: "kube-api-access-g4nzm") pod "206f5ad4-5290-4ac9-86c2-b4e945849790" (UID: "206f5ad4-5290-4ac9-86c2-b4e945849790"). InnerVolumeSpecName "kube-api-access-g4nzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:16:32 crc kubenswrapper[4806]: I1204 05:16:32.988163 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/206f5ad4-5290-4ac9-86c2-b4e945849790-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 05:16:32 crc kubenswrapper[4806]: I1204 05:16:32.988367 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4nzm\" (UniqueName: \"kubernetes.io/projected/206f5ad4-5290-4ac9-86c2-b4e945849790-kube-api-access-g4nzm\") on node \"crc\" DevicePath \"\"" Dec 04 05:16:32 crc kubenswrapper[4806]: I1204 05:16:32.988459 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/206f5ad4-5290-4ac9-86c2-b4e945849790-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 05:16:33 crc kubenswrapper[4806]: I1204 05:16:33.341151 4806 generic.go:334] "Generic (PLEG): container finished" podID="206f5ad4-5290-4ac9-86c2-b4e945849790" containerID="3e080fc639e263395b2252ec5afdcb24a3e6717ffe0af47960f0602921ec3519" exitCode=0 Dec 04 05:16:33 crc kubenswrapper[4806]: I1204 05:16:33.341194 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vvqv" event={"ID":"206f5ad4-5290-4ac9-86c2-b4e945849790","Type":"ContainerDied","Data":"3e080fc639e263395b2252ec5afdcb24a3e6717ffe0af47960f0602921ec3519"} Dec 04 05:16:33 crc kubenswrapper[4806]: I1204 05:16:33.341277 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vvqv" event={"ID":"206f5ad4-5290-4ac9-86c2-b4e945849790","Type":"ContainerDied","Data":"d5a2befc60b57b40209ddf179cbae34ebecea2ec37229cb789d5631a5dce1ef0"} Dec 04 05:16:33 crc kubenswrapper[4806]: I1204 05:16:33.341280 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4vvqv" Dec 04 05:16:33 crc kubenswrapper[4806]: I1204 05:16:33.341303 4806 scope.go:117] "RemoveContainer" containerID="3e080fc639e263395b2252ec5afdcb24a3e6717ffe0af47960f0602921ec3519" Dec 04 05:16:33 crc kubenswrapper[4806]: I1204 05:16:33.378671 4806 scope.go:117] "RemoveContainer" containerID="9a2cfd286a0070df72755905336e44e402924e05f777ff733f967a10c8bfb2a5" Dec 04 05:16:33 crc kubenswrapper[4806]: I1204 05:16:33.391010 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4vvqv"] Dec 04 05:16:33 crc kubenswrapper[4806]: I1204 05:16:33.406813 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4vvqv"] Dec 04 05:16:33 crc kubenswrapper[4806]: I1204 05:16:33.408108 4806 scope.go:117] "RemoveContainer" containerID="05bca03557fd83437b470a8caec95e2382ecbb2642c1ba4fba9dbb261bd05a1e" Dec 04 05:16:33 crc kubenswrapper[4806]: I1204 05:16:33.449341 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="206f5ad4-5290-4ac9-86c2-b4e945849790" path="/var/lib/kubelet/pods/206f5ad4-5290-4ac9-86c2-b4e945849790/volumes" Dec 04 05:16:33 crc kubenswrapper[4806]: I1204 05:16:33.471070 4806 scope.go:117] "RemoveContainer" containerID="3e080fc639e263395b2252ec5afdcb24a3e6717ffe0af47960f0602921ec3519" Dec 04 05:16:33 crc kubenswrapper[4806]: E1204 05:16:33.471492 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e080fc639e263395b2252ec5afdcb24a3e6717ffe0af47960f0602921ec3519\": container with ID starting with 3e080fc639e263395b2252ec5afdcb24a3e6717ffe0af47960f0602921ec3519 not found: ID does not exist" containerID="3e080fc639e263395b2252ec5afdcb24a3e6717ffe0af47960f0602921ec3519" Dec 04 05:16:33 crc kubenswrapper[4806]: I1204 05:16:33.471529 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e080fc639e263395b2252ec5afdcb24a3e6717ffe0af47960f0602921ec3519"} err="failed to get container status \"3e080fc639e263395b2252ec5afdcb24a3e6717ffe0af47960f0602921ec3519\": rpc error: code = NotFound desc = could not find container \"3e080fc639e263395b2252ec5afdcb24a3e6717ffe0af47960f0602921ec3519\": container with ID starting with 3e080fc639e263395b2252ec5afdcb24a3e6717ffe0af47960f0602921ec3519 not found: ID does not exist" Dec 04 05:16:33 crc kubenswrapper[4806]: I1204 05:16:33.471554 4806 scope.go:117] "RemoveContainer" containerID="9a2cfd286a0070df72755905336e44e402924e05f777ff733f967a10c8bfb2a5" Dec 04 05:16:33 crc kubenswrapper[4806]: E1204 05:16:33.471960 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a2cfd286a0070df72755905336e44e402924e05f777ff733f967a10c8bfb2a5\": container with ID starting with 9a2cfd286a0070df72755905336e44e402924e05f777ff733f967a10c8bfb2a5 not found: ID does not exist" containerID="9a2cfd286a0070df72755905336e44e402924e05f777ff733f967a10c8bfb2a5" Dec 04 05:16:33 crc kubenswrapper[4806]: I1204 05:16:33.471988 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a2cfd286a0070df72755905336e44e402924e05f777ff733f967a10c8bfb2a5"} err="failed to get container status \"9a2cfd286a0070df72755905336e44e402924e05f777ff733f967a10c8bfb2a5\": rpc error: code = NotFound desc = could not find container \"9a2cfd286a0070df72755905336e44e402924e05f777ff733f967a10c8bfb2a5\": container with ID starting with 9a2cfd286a0070df72755905336e44e402924e05f777ff733f967a10c8bfb2a5 not found: ID does not exist" Dec 04 05:16:33 crc kubenswrapper[4806]: I1204 05:16:33.472007 4806 scope.go:117] "RemoveContainer" containerID="05bca03557fd83437b470a8caec95e2382ecbb2642c1ba4fba9dbb261bd05a1e" Dec 04 05:16:33 crc kubenswrapper[4806]: E1204 05:16:33.472296 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05bca03557fd83437b470a8caec95e2382ecbb2642c1ba4fba9dbb261bd05a1e\": container with ID starting with 05bca03557fd83437b470a8caec95e2382ecbb2642c1ba4fba9dbb261bd05a1e not found: ID does not exist" containerID="05bca03557fd83437b470a8caec95e2382ecbb2642c1ba4fba9dbb261bd05a1e" Dec 04 05:16:33 crc kubenswrapper[4806]: I1204 05:16:33.472323 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05bca03557fd83437b470a8caec95e2382ecbb2642c1ba4fba9dbb261bd05a1e"} err="failed to get container status \"05bca03557fd83437b470a8caec95e2382ecbb2642c1ba4fba9dbb261bd05a1e\": rpc error: code = NotFound desc = could not find container \"05bca03557fd83437b470a8caec95e2382ecbb2642c1ba4fba9dbb261bd05a1e\": container with ID starting with 05bca03557fd83437b470a8caec95e2382ecbb2642c1ba4fba9dbb261bd05a1e not found: ID does not exist" Dec 04 05:16:38 crc kubenswrapper[4806]: I1204 05:16:38.621901 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-6l44d_ed32bab6-fee8-43d7-a157-9e8d5d17b98a/cert-manager-controller/0.log" Dec 04 05:16:38 crc kubenswrapper[4806]: I1204 05:16:38.810151 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-8vl9s_65faa3e8-8810-4a0e-a3ec-bb2b4d6283fb/cert-manager-cainjector/0.log" Dec 04 05:16:38 crc kubenswrapper[4806]: I1204 05:16:38.926715 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-dzbcz_7934a456-a3ec-4a88-8857-17d7f7db5837/cert-manager-webhook/0.log" Dec 04 05:16:53 crc kubenswrapper[4806]: I1204 05:16:53.393543 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-572xx_b0e181e1-d1fa-460f-963f-cbd235500b2e/nmstate-console-plugin/0.log" Dec 04 05:16:53 crc kubenswrapper[4806]: I1204 05:16:53.607342 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-pdg7x_843789a7-01bb-42c0-9f30-aa52af788147/nmstate-handler/0.log" Dec 04 05:16:53 crc kubenswrapper[4806]: I1204 05:16:53.630323 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-dlkzr_7b294b89-f409-40b7-8ba4-72e53da24a6f/kube-rbac-proxy/0.log" Dec 04 05:16:53 crc kubenswrapper[4806]: I1204 05:16:53.694449 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-dlkzr_7b294b89-f409-40b7-8ba4-72e53da24a6f/nmstate-metrics/0.log" Dec 04 05:16:53 crc kubenswrapper[4806]: I1204 05:16:53.906930 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-zz62j_93c65cc3-ea7f-4f77-af33-8acdcd03432c/nmstate-operator/0.log" Dec 04 05:16:53 crc kubenswrapper[4806]: I1204 05:16:53.908535 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-q9x68_eae97ef2-3aeb-4303-b26f-0f84a75abda0/nmstate-webhook/0.log" Dec 04 05:16:57 crc kubenswrapper[4806]: I1204 05:16:57.047699 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:16:57 crc kubenswrapper[4806]: I1204 05:16:57.049381 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:16:57 crc kubenswrapper[4806]: I1204 05:16:57.049535 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 05:16:57 crc kubenswrapper[4806]: I1204 05:16:57.050427 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 05:16:57 crc kubenswrapper[4806]: I1204 05:16:57.050586 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" gracePeriod=600 Dec 04 05:16:57 crc kubenswrapper[4806]: E1204 05:16:57.181780 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:16:57 crc kubenswrapper[4806]: I1204 05:16:57.588207 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" exitCode=0 Dec 04 05:16:57 crc kubenswrapper[4806]: I1204 05:16:57.588265 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459"} Dec 04 05:16:57 crc kubenswrapper[4806]: I1204 05:16:57.588306 4806 scope.go:117] "RemoveContainer" containerID="a1e6c4d9918f0691a954f6d7dd837700fe9fc560ddcdf3bc751c5b70d3e912e2" Dec 04 05:16:57 crc kubenswrapper[4806]: I1204 05:16:57.589151 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:16:57 crc kubenswrapper[4806]: E1204 05:16:57.589646 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:17:08 crc kubenswrapper[4806]: I1204 05:17:08.424777 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:17:08 crc kubenswrapper[4806]: E1204 05:17:08.426896 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:17:09 crc kubenswrapper[4806]: I1204 05:17:09.075892 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-qq2x8_d80ba885-bdec-472c-b9eb-23928eeb6ab1/kube-rbac-proxy/0.log" Dec 04 05:17:09 crc kubenswrapper[4806]: I1204 05:17:09.192837 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-qq2x8_d80ba885-bdec-472c-b9eb-23928eeb6ab1/controller/0.log" Dec 04 05:17:09 crc kubenswrapper[4806]: I1204 05:17:09.325467 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-frr-files/0.log" Dec 04 05:17:09 crc kubenswrapper[4806]: I1204 05:17:09.628842 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-metrics/0.log" Dec 04 05:17:09 crc kubenswrapper[4806]: I1204 05:17:09.646185 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-reloader/0.log" Dec 04 05:17:09 crc kubenswrapper[4806]: I1204 05:17:09.658357 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-frr-files/0.log" Dec 04 05:17:09 crc kubenswrapper[4806]: I1204 05:17:09.670451 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-reloader/0.log" Dec 04 05:17:09 crc kubenswrapper[4806]: I1204 05:17:09.838683 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-reloader/0.log" Dec 04 05:17:09 crc kubenswrapper[4806]: I1204 05:17:09.843876 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-metrics/0.log" Dec 04 05:17:09 crc kubenswrapper[4806]: I1204 05:17:09.859978 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-frr-files/0.log" Dec 04 05:17:09 crc kubenswrapper[4806]: I1204 05:17:09.913973 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-metrics/0.log" Dec 04 05:17:10 crc kubenswrapper[4806]: I1204 05:17:10.089598 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-frr-files/0.log" Dec 04 05:17:10 crc kubenswrapper[4806]: I1204 05:17:10.117941 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-reloader/0.log" Dec 04 05:17:10 crc kubenswrapper[4806]: I1204 05:17:10.140142 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-metrics/0.log" Dec 04 05:17:10 crc kubenswrapper[4806]: I1204 05:17:10.152627 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/controller/0.log" Dec 04 05:17:10 crc kubenswrapper[4806]: I1204 05:17:10.349964 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/frr-metrics/0.log" Dec 04 05:17:10 crc kubenswrapper[4806]: I1204 05:17:10.411473 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/kube-rbac-proxy/0.log" Dec 04 05:17:10 crc kubenswrapper[4806]: I1204 05:17:10.462408 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/kube-rbac-proxy-frr/0.log" Dec 04 05:17:10 crc kubenswrapper[4806]: I1204 05:17:10.975908 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/reloader/0.log" Dec 04 05:17:11 crc kubenswrapper[4806]: I1204 05:17:11.042918 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-wh624_fac17a35-6b82-4397-855c-a9d2b27bcca6/frr-k8s-webhook-server/0.log" Dec 04 05:17:11 crc kubenswrapper[4806]: I1204 05:17:11.241900 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5b99fcbdc-n6szh_09d57bc9-a5d5-4848-9cbd-12ce41ca2b78/manager/0.log" Dec 04 05:17:11 crc kubenswrapper[4806]: I1204 05:17:11.537224 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-58dfcfbb98-7prrp_134cc417-ef0e-4cb7-90e5-5a8f55466f6a/webhook-server/0.log" Dec 04 05:17:11 crc kubenswrapper[4806]: I1204 05:17:11.652002 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-g44sf_e653dfc5-42b0-4ae3-a622-b64b556d3ccc/kube-rbac-proxy/0.log" Dec 04 05:17:11 crc kubenswrapper[4806]: I1204 05:17:11.811417 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/frr/0.log" Dec 04 05:17:12 crc kubenswrapper[4806]: I1204 05:17:12.164772 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-g44sf_e653dfc5-42b0-4ae3-a622-b64b556d3ccc/speaker/0.log" Dec 04 05:17:21 crc kubenswrapper[4806]: I1204 05:17:21.425867 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:17:21 crc kubenswrapper[4806]: E1204 05:17:21.427682 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:17:26 crc kubenswrapper[4806]: I1204 05:17:26.887957 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b_8981d20c-af12-4497-abc8-f17930ec12c1/util/0.log" Dec 04 05:17:27 crc kubenswrapper[4806]: I1204 05:17:27.475528 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b_8981d20c-af12-4497-abc8-f17930ec12c1/util/0.log" Dec 04 05:17:27 crc kubenswrapper[4806]: I1204 05:17:27.476249 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b_8981d20c-af12-4497-abc8-f17930ec12c1/pull/0.log" Dec 04 05:17:27 crc kubenswrapper[4806]: I1204 05:17:27.488985 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b_8981d20c-af12-4497-abc8-f17930ec12c1/pull/0.log" Dec 04 05:17:27 crc kubenswrapper[4806]: I1204 05:17:27.779834 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b_8981d20c-af12-4497-abc8-f17930ec12c1/util/0.log" Dec 04 05:17:27 crc kubenswrapper[4806]: I1204 05:17:27.790443 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b_8981d20c-af12-4497-abc8-f17930ec12c1/pull/0.log" Dec 04 05:17:27 crc kubenswrapper[4806]: I1204 05:17:27.810797 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b_8981d20c-af12-4497-abc8-f17930ec12c1/extract/0.log" Dec 04 05:17:28 crc kubenswrapper[4806]: I1204 05:17:28.042216 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4_205b1465-25a7-43a5-9449-78baf9a87116/util/0.log" Dec 04 05:17:28 crc kubenswrapper[4806]: I1204 05:17:28.992843 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4_205b1465-25a7-43a5-9449-78baf9a87116/pull/0.log" Dec 04 05:17:29 crc kubenswrapper[4806]: I1204 05:17:29.010901 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4_205b1465-25a7-43a5-9449-78baf9a87116/util/0.log" Dec 04 05:17:29 crc kubenswrapper[4806]: I1204 05:17:29.033096 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4_205b1465-25a7-43a5-9449-78baf9a87116/pull/0.log" Dec 04 05:17:29 crc kubenswrapper[4806]: I1204 05:17:29.311858 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4_205b1465-25a7-43a5-9449-78baf9a87116/util/0.log" Dec 04 05:17:29 crc kubenswrapper[4806]: I1204 05:17:29.337175 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4_205b1465-25a7-43a5-9449-78baf9a87116/pull/0.log" Dec 04 05:17:29 crc kubenswrapper[4806]: I1204 05:17:29.337345 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4_205b1465-25a7-43a5-9449-78baf9a87116/extract/0.log" Dec 04 05:17:29 crc kubenswrapper[4806]: I1204 05:17:29.590538 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mkm6b_c3bcad8b-9a77-4d44-9bde-c3890fe6427c/extract-utilities/0.log" Dec 04 05:17:29 crc kubenswrapper[4806]: I1204 05:17:29.805901 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mkm6b_c3bcad8b-9a77-4d44-9bde-c3890fe6427c/extract-content/0.log" Dec 04 05:17:29 crc kubenswrapper[4806]: I1204 05:17:29.817619 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mkm6b_c3bcad8b-9a77-4d44-9bde-c3890fe6427c/extract-utilities/0.log" Dec 04 05:17:29 crc kubenswrapper[4806]: I1204 05:17:29.836665 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mkm6b_c3bcad8b-9a77-4d44-9bde-c3890fe6427c/extract-content/0.log" Dec 04 05:17:30 crc kubenswrapper[4806]: I1204 05:17:30.031360 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mkm6b_c3bcad8b-9a77-4d44-9bde-c3890fe6427c/extract-content/0.log" Dec 04 05:17:30 crc kubenswrapper[4806]: I1204 05:17:30.038287 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mkm6b_c3bcad8b-9a77-4d44-9bde-c3890fe6427c/extract-utilities/0.log" Dec 04 05:17:30 crc kubenswrapper[4806]: I1204 05:17:30.738958 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2bzpg_0bd6d288-1d0e-49a2-b9de-5437c5d22356/extract-utilities/0.log" Dec 04 05:17:30 crc kubenswrapper[4806]: I1204 05:17:30.746580 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mkm6b_c3bcad8b-9a77-4d44-9bde-c3890fe6427c/registry-server/0.log" Dec 04 05:17:30 crc kubenswrapper[4806]: I1204 05:17:30.990015 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2bzpg_0bd6d288-1d0e-49a2-b9de-5437c5d22356/extract-utilities/0.log" Dec 04 05:17:31 crc kubenswrapper[4806]: I1204 05:17:31.001647 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2bzpg_0bd6d288-1d0e-49a2-b9de-5437c5d22356/extract-content/0.log" Dec 04 05:17:31 crc kubenswrapper[4806]: I1204 05:17:31.019672 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2bzpg_0bd6d288-1d0e-49a2-b9de-5437c5d22356/extract-content/0.log" Dec 04 05:17:31 crc kubenswrapper[4806]: I1204 05:17:31.194611 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2bzpg_0bd6d288-1d0e-49a2-b9de-5437c5d22356/extract-utilities/0.log" Dec 04 05:17:31 crc kubenswrapper[4806]: I1204 05:17:31.229232 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2bzpg_0bd6d288-1d0e-49a2-b9de-5437c5d22356/extract-content/0.log" Dec 04 05:17:31 crc kubenswrapper[4806]: I1204 05:17:31.342243 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9rvth_60ce8b7c-5cf6-452c-835e-9ab84a66aa65/marketplace-operator/3.log" Dec 04 05:17:31 crc kubenswrapper[4806]: I1204 05:17:31.438104 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9rvth_60ce8b7c-5cf6-452c-835e-9ab84a66aa65/marketplace-operator/2.log" Dec 04 05:17:31 crc kubenswrapper[4806]: I1204 05:17:31.546410 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kgr4l_7bf8469a-647b-4b61-ab6c-89bbcc436ab9/extract-utilities/0.log" Dec 04 05:17:31 crc kubenswrapper[4806]: I1204 05:17:31.949190 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kgr4l_7bf8469a-647b-4b61-ab6c-89bbcc436ab9/extract-utilities/0.log" Dec 04 05:17:31 crc kubenswrapper[4806]: I1204 05:17:31.975706 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2bzpg_0bd6d288-1d0e-49a2-b9de-5437c5d22356/registry-server/0.log" Dec 04 05:17:31 crc kubenswrapper[4806]: I1204 05:17:31.985852 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kgr4l_7bf8469a-647b-4b61-ab6c-89bbcc436ab9/extract-content/0.log" Dec 04 05:17:32 crc kubenswrapper[4806]: I1204 05:17:32.012698 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kgr4l_7bf8469a-647b-4b61-ab6c-89bbcc436ab9/extract-content/0.log" Dec 04 05:17:32 crc kubenswrapper[4806]: I1204 05:17:32.128248 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kgr4l_7bf8469a-647b-4b61-ab6c-89bbcc436ab9/extract-utilities/0.log" Dec 04 05:17:32 crc kubenswrapper[4806]: I1204 05:17:32.128625 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kgr4l_7bf8469a-647b-4b61-ab6c-89bbcc436ab9/extract-content/0.log" Dec 04 05:17:32 crc kubenswrapper[4806]: I1204 05:17:32.215500 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrtjm_242a6296-78ef-431a-8c26-522e66e5296c/extract-utilities/0.log" Dec 04 05:17:32 crc kubenswrapper[4806]: I1204 05:17:32.384811 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kgr4l_7bf8469a-647b-4b61-ab6c-89bbcc436ab9/registry-server/0.log" Dec 04 05:17:32 crc kubenswrapper[4806]: I1204 05:17:32.463535 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrtjm_242a6296-78ef-431a-8c26-522e66e5296c/extract-content/0.log" Dec 04 05:17:32 crc kubenswrapper[4806]: I1204 05:17:32.472954 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrtjm_242a6296-78ef-431a-8c26-522e66e5296c/extract-content/0.log" Dec 04 05:17:32 crc kubenswrapper[4806]: I1204 05:17:32.495104 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrtjm_242a6296-78ef-431a-8c26-522e66e5296c/extract-utilities/0.log" Dec 04 05:17:32 crc kubenswrapper[4806]: I1204 05:17:32.596187 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrtjm_242a6296-78ef-431a-8c26-522e66e5296c/extract-utilities/0.log" Dec 04 05:17:32 crc kubenswrapper[4806]: I1204 05:17:32.716754 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrtjm_242a6296-78ef-431a-8c26-522e66e5296c/extract-content/0.log" Dec 04 05:17:33 crc kubenswrapper[4806]: I1204 05:17:33.188083 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrtjm_242a6296-78ef-431a-8c26-522e66e5296c/registry-server/0.log" Dec 04 05:17:35 crc kubenswrapper[4806]: I1204 05:17:35.430960 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:17:35 crc kubenswrapper[4806]: E1204 05:17:35.431541 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:17:50 crc kubenswrapper[4806]: I1204 05:17:50.425132 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:17:50 crc kubenswrapper[4806]: E1204 05:17:50.426070 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:18:01 crc kubenswrapper[4806]: I1204 05:18:01.430891 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:18:01 crc kubenswrapper[4806]: E1204 05:18:01.431826 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:18:14 crc kubenswrapper[4806]: I1204 05:18:14.423448 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:18:14 crc kubenswrapper[4806]: E1204 05:18:14.424251 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:18:27 crc kubenswrapper[4806]: I1204 05:18:27.425598 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:18:27 crc kubenswrapper[4806]: E1204 05:18:27.426313 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:18:38 crc kubenswrapper[4806]: I1204 05:18:38.423626 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:18:38 crc kubenswrapper[4806]: E1204 05:18:38.424527 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:18:49 crc kubenswrapper[4806]: I1204 05:18:49.424556 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:18:49 crc kubenswrapper[4806]: E1204 05:18:49.425575 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:19:01 crc kubenswrapper[4806]: I1204 05:19:01.423949 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:19:01 crc kubenswrapper[4806]: E1204 05:19:01.424859 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:19:14 crc kubenswrapper[4806]: I1204 05:19:14.423585 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:19:14 crc kubenswrapper[4806]: E1204 05:19:14.424462 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:19:27 crc kubenswrapper[4806]: I1204 05:19:27.423371 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:19:27 crc kubenswrapper[4806]: E1204 05:19:27.424093 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:19:38 crc kubenswrapper[4806]: I1204 05:19:38.423675 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:19:38 crc kubenswrapper[4806]: E1204 05:19:38.424732 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:19:50 crc kubenswrapper[4806]: I1204 05:19:50.423376 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:19:50 crc kubenswrapper[4806]: E1204 05:19:50.424085 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:19:57 crc kubenswrapper[4806]: I1204 05:19:57.353743 4806 generic.go:334] "Generic (PLEG): container finished" podID="1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49" containerID="a47a02f05a9c602c361615d3351a2df5e5eb547a262626b51245aab28658722d" exitCode=0 Dec 04 05:19:57 crc kubenswrapper[4806]: I1204 05:19:57.353814 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zx5tm/must-gather-mxl86" event={"ID":"1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49","Type":"ContainerDied","Data":"a47a02f05a9c602c361615d3351a2df5e5eb547a262626b51245aab28658722d"} Dec 04 05:19:57 crc kubenswrapper[4806]: I1204 05:19:57.355312 4806 scope.go:117] "RemoveContainer" containerID="a47a02f05a9c602c361615d3351a2df5e5eb547a262626b51245aab28658722d" Dec 04 05:19:57 crc kubenswrapper[4806]: I1204 05:19:57.956668 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zx5tm_must-gather-mxl86_1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49/gather/0.log" Dec 04 05:20:03 crc kubenswrapper[4806]: I1204 05:20:03.424313 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:20:03 crc kubenswrapper[4806]: E1204 05:20:03.425067 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:20:07 crc kubenswrapper[4806]: I1204 05:20:07.306554 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zx5tm/must-gather-mxl86"] Dec 04 05:20:07 crc kubenswrapper[4806]: I1204 05:20:07.307445 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-zx5tm/must-gather-mxl86" podUID="1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49" containerName="copy" containerID="cri-o://471251cb2cdde3685841d548b1587e8e8144feed36a229eea8646086651bfa91" gracePeriod=2 Dec 04 05:20:07 crc kubenswrapper[4806]: I1204 05:20:07.339764 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zx5tm/must-gather-mxl86"] Dec 04 05:20:07 crc kubenswrapper[4806]: I1204 05:20:07.447342 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zx5tm_must-gather-mxl86_1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49/copy/0.log" Dec 04 05:20:07 crc kubenswrapper[4806]: I1204 05:20:07.448414 4806 generic.go:334] "Generic (PLEG): container finished" podID="1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49" containerID="471251cb2cdde3685841d548b1587e8e8144feed36a229eea8646086651bfa91" exitCode=143 Dec 04 05:20:07 crc kubenswrapper[4806]: I1204 05:20:07.758092 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zx5tm_must-gather-mxl86_1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49/copy/0.log" Dec 04 05:20:07 crc kubenswrapper[4806]: I1204 05:20:07.758863 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx5tm/must-gather-mxl86" Dec 04 05:20:07 crc kubenswrapper[4806]: I1204 05:20:07.866808 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49-must-gather-output\") pod \"1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49\" (UID: \"1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49\") " Dec 04 05:20:07 crc kubenswrapper[4806]: I1204 05:20:07.866973 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkbvc\" (UniqueName: \"kubernetes.io/projected/1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49-kube-api-access-qkbvc\") pod \"1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49\" (UID: \"1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49\") " Dec 04 05:20:07 crc kubenswrapper[4806]: I1204 05:20:07.872974 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49-kube-api-access-qkbvc" (OuterVolumeSpecName: "kube-api-access-qkbvc") pod "1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49" (UID: "1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49"). InnerVolumeSpecName "kube-api-access-qkbvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:20:07 crc kubenswrapper[4806]: I1204 05:20:07.971257 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkbvc\" (UniqueName: \"kubernetes.io/projected/1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49-kube-api-access-qkbvc\") on node \"crc\" DevicePath \"\"" Dec 04 05:20:08 crc kubenswrapper[4806]: I1204 05:20:08.079351 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49" (UID: "1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:20:08 crc kubenswrapper[4806]: I1204 05:20:08.175511 4806 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 04 05:20:08 crc kubenswrapper[4806]: I1204 05:20:08.463444 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zx5tm_must-gather-mxl86_1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49/copy/0.log" Dec 04 05:20:08 crc kubenswrapper[4806]: I1204 05:20:08.464127 4806 scope.go:117] "RemoveContainer" containerID="471251cb2cdde3685841d548b1587e8e8144feed36a229eea8646086651bfa91" Dec 04 05:20:08 crc kubenswrapper[4806]: I1204 05:20:08.464297 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zx5tm/must-gather-mxl86" Dec 04 05:20:08 crc kubenswrapper[4806]: I1204 05:20:08.494643 4806 scope.go:117] "RemoveContainer" containerID="a47a02f05a9c602c361615d3351a2df5e5eb547a262626b51245aab28658722d" Dec 04 05:20:09 crc kubenswrapper[4806]: I1204 05:20:09.433853 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49" path="/var/lib/kubelet/pods/1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49/volumes" Dec 04 05:20:15 crc kubenswrapper[4806]: I1204 05:20:15.434409 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:20:15 crc kubenswrapper[4806]: E1204 05:20:15.435070 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:20:30 crc kubenswrapper[4806]: I1204 05:20:30.423752 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:20:30 crc kubenswrapper[4806]: E1204 05:20:30.424540 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:20:43 crc kubenswrapper[4806]: I1204 05:20:43.423760 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:20:43 crc kubenswrapper[4806]: E1204 05:20:43.424438 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:20:56 crc kubenswrapper[4806]: I1204 05:20:56.424123 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:20:56 crc kubenswrapper[4806]: E1204 05:20:56.424840 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:21:03 crc kubenswrapper[4806]: I1204 05:21:03.286525 4806 scope.go:117] "RemoveContainer" containerID="333e066ce103c754b931cd382b76646a6ae9b277085c829f90d63556845c6290" Dec 04 05:21:10 crc kubenswrapper[4806]: I1204 05:21:10.424007 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:21:10 crc kubenswrapper[4806]: E1204 05:21:10.424874 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:21:24 crc kubenswrapper[4806]: I1204 05:21:24.423966 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:21:24 crc kubenswrapper[4806]: E1204 05:21:24.424911 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:21:36 crc kubenswrapper[4806]: I1204 05:21:36.424821 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:21:36 crc kubenswrapper[4806]: E1204 05:21:36.426157 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:21:47 crc kubenswrapper[4806]: I1204 05:21:47.426613 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:21:47 crc kubenswrapper[4806]: E1204 05:21:47.428785 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:22:02 crc kubenswrapper[4806]: I1204 05:22:02.424028 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:22:02 crc kubenswrapper[4806]: I1204 05:22:02.789191 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"a025afa4e0f2bafc160c671d65ca45902522db3c999cd6d04dd15c40321dcf4a"} Dec 04 05:23:02 crc kubenswrapper[4806]: I1204 05:23:02.871695 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-dz69k/must-gather-zmxrl"] Dec 04 05:23:02 crc kubenswrapper[4806]: E1204 05:23:02.885477 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49" containerName="gather" Dec 04 05:23:02 crc kubenswrapper[4806]: I1204 05:23:02.885534 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49" containerName="gather" Dec 04 05:23:02 crc kubenswrapper[4806]: E1204 05:23:02.885593 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="206f5ad4-5290-4ac9-86c2-b4e945849790" containerName="extract-utilities" Dec 04 05:23:02 crc kubenswrapper[4806]: I1204 05:23:02.885608 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="206f5ad4-5290-4ac9-86c2-b4e945849790" containerName="extract-utilities" Dec 04 05:23:02 crc kubenswrapper[4806]: E1204 05:23:02.885644 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="206f5ad4-5290-4ac9-86c2-b4e945849790" containerName="extract-content" Dec 04 05:23:02 crc kubenswrapper[4806]: I1204 05:23:02.885659 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="206f5ad4-5290-4ac9-86c2-b4e945849790" containerName="extract-content" Dec 04 05:23:02 crc kubenswrapper[4806]: E1204 05:23:02.885679 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49" containerName="copy" Dec 04 05:23:02 crc kubenswrapper[4806]: I1204 05:23:02.885687 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49" containerName="copy" Dec 04 05:23:02 crc kubenswrapper[4806]: E1204 05:23:02.885724 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="206f5ad4-5290-4ac9-86c2-b4e945849790" containerName="registry-server" Dec 04 05:23:02 crc kubenswrapper[4806]: I1204 05:23:02.885733 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="206f5ad4-5290-4ac9-86c2-b4e945849790" containerName="registry-server" Dec 04 05:23:02 crc kubenswrapper[4806]: I1204 05:23:02.886264 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49" containerName="copy" Dec 04 05:23:02 crc kubenswrapper[4806]: I1204 05:23:02.886340 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e97e1aa-f84c-4e2d-b6bd-471d1dcb4d49" containerName="gather" Dec 04 05:23:02 crc kubenswrapper[4806]: I1204 05:23:02.886353 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="206f5ad4-5290-4ac9-86c2-b4e945849790" containerName="registry-server" Dec 04 05:23:02 crc kubenswrapper[4806]: I1204 05:23:02.913367 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dz69k/must-gather-zmxrl" Dec 04 05:23:02 crc kubenswrapper[4806]: I1204 05:23:02.935629 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-dz69k"/"openshift-service-ca.crt" Dec 04 05:23:02 crc kubenswrapper[4806]: I1204 05:23:02.982976 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-dz69k/must-gather-zmxrl"] Dec 04 05:23:02 crc kubenswrapper[4806]: I1204 05:23:02.985180 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-dz69k"/"kube-root-ca.crt" Dec 04 05:23:03 crc kubenswrapper[4806]: I1204 05:23:03.078179 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ed7a3f69-8963-405c-b661-3763e87d2a69-must-gather-output\") pod \"must-gather-zmxrl\" (UID: \"ed7a3f69-8963-405c-b661-3763e87d2a69\") " pod="openshift-must-gather-dz69k/must-gather-zmxrl" Dec 04 05:23:03 crc kubenswrapper[4806]: I1204 05:23:03.078743 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzllh\" (UniqueName: \"kubernetes.io/projected/ed7a3f69-8963-405c-b661-3763e87d2a69-kube-api-access-rzllh\") pod \"must-gather-zmxrl\" (UID: \"ed7a3f69-8963-405c-b661-3763e87d2a69\") " pod="openshift-must-gather-dz69k/must-gather-zmxrl" Dec 04 05:23:03 crc kubenswrapper[4806]: I1204 05:23:03.180249 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzllh\" (UniqueName: \"kubernetes.io/projected/ed7a3f69-8963-405c-b661-3763e87d2a69-kube-api-access-rzllh\") pod \"must-gather-zmxrl\" (UID: \"ed7a3f69-8963-405c-b661-3763e87d2a69\") " pod="openshift-must-gather-dz69k/must-gather-zmxrl" Dec 04 05:23:03 crc kubenswrapper[4806]: I1204 05:23:03.180372 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ed7a3f69-8963-405c-b661-3763e87d2a69-must-gather-output\") pod \"must-gather-zmxrl\" (UID: \"ed7a3f69-8963-405c-b661-3763e87d2a69\") " pod="openshift-must-gather-dz69k/must-gather-zmxrl" Dec 04 05:23:03 crc kubenswrapper[4806]: I1204 05:23:03.180880 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ed7a3f69-8963-405c-b661-3763e87d2a69-must-gather-output\") pod \"must-gather-zmxrl\" (UID: \"ed7a3f69-8963-405c-b661-3763e87d2a69\") " pod="openshift-must-gather-dz69k/must-gather-zmxrl" Dec 04 05:23:03 crc kubenswrapper[4806]: I1204 05:23:03.198529 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzllh\" (UniqueName: \"kubernetes.io/projected/ed7a3f69-8963-405c-b661-3763e87d2a69-kube-api-access-rzllh\") pod \"must-gather-zmxrl\" (UID: \"ed7a3f69-8963-405c-b661-3763e87d2a69\") " pod="openshift-must-gather-dz69k/must-gather-zmxrl" Dec 04 05:23:03 crc kubenswrapper[4806]: I1204 05:23:03.306360 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dz69k/must-gather-zmxrl" Dec 04 05:23:03 crc kubenswrapper[4806]: I1204 05:23:03.838843 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-dz69k/must-gather-zmxrl"] Dec 04 05:23:04 crc kubenswrapper[4806]: I1204 05:23:04.368700 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dz69k/must-gather-zmxrl" event={"ID":"ed7a3f69-8963-405c-b661-3763e87d2a69","Type":"ContainerStarted","Data":"cc336c30daa805af38a23f52cfb34297e2da0cc24d084abde460b4c4a02d7b50"} Dec 04 05:23:04 crc kubenswrapper[4806]: I1204 05:23:04.369018 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dz69k/must-gather-zmxrl" event={"ID":"ed7a3f69-8963-405c-b661-3763e87d2a69","Type":"ContainerStarted","Data":"1a9680f835af0cd6381d86063655db8c2a0e50de145e3681dbb489118ea33c70"} Dec 04 05:23:05 crc kubenswrapper[4806]: I1204 05:23:05.378442 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dz69k/must-gather-zmxrl" event={"ID":"ed7a3f69-8963-405c-b661-3763e87d2a69","Type":"ContainerStarted","Data":"8cac059cac8ee864d72474a8e6d0307b6668bead6bdf640dfcfd5c0d86f58919"} Dec 04 05:23:05 crc kubenswrapper[4806]: I1204 05:23:05.397903 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-dz69k/must-gather-zmxrl" podStartSLOduration=3.397880227 podStartE2EDuration="3.397880227s" podCreationTimestamp="2025-12-04 05:23:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 05:23:05.391469873 +0000 UTC m=+5300.249982811" watchObservedRunningTime="2025-12-04 05:23:05.397880227 +0000 UTC m=+5300.256393155" Dec 04 05:23:08 crc kubenswrapper[4806]: I1204 05:23:08.047449 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-dz69k/crc-debug-657c4"] Dec 04 05:23:08 crc kubenswrapper[4806]: I1204 05:23:08.049737 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dz69k/crc-debug-657c4" Dec 04 05:23:08 crc kubenswrapper[4806]: I1204 05:23:08.056914 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-dz69k"/"default-dockercfg-fdnwj" Dec 04 05:23:08 crc kubenswrapper[4806]: I1204 05:23:08.175507 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxn5t\" (UniqueName: \"kubernetes.io/projected/5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2-kube-api-access-fxn5t\") pod \"crc-debug-657c4\" (UID: \"5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2\") " pod="openshift-must-gather-dz69k/crc-debug-657c4" Dec 04 05:23:08 crc kubenswrapper[4806]: I1204 05:23:08.175657 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2-host\") pod \"crc-debug-657c4\" (UID: \"5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2\") " pod="openshift-must-gather-dz69k/crc-debug-657c4" Dec 04 05:23:08 crc kubenswrapper[4806]: I1204 05:23:08.277554 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxn5t\" (UniqueName: \"kubernetes.io/projected/5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2-kube-api-access-fxn5t\") pod \"crc-debug-657c4\" (UID: \"5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2\") " pod="openshift-must-gather-dz69k/crc-debug-657c4" Dec 04 05:23:08 crc kubenswrapper[4806]: I1204 05:23:08.277659 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2-host\") pod \"crc-debug-657c4\" (UID: \"5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2\") " pod="openshift-must-gather-dz69k/crc-debug-657c4" Dec 04 05:23:08 crc kubenswrapper[4806]: I1204 05:23:08.277775 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2-host\") pod \"crc-debug-657c4\" (UID: \"5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2\") " pod="openshift-must-gather-dz69k/crc-debug-657c4" Dec 04 05:23:08 crc kubenswrapper[4806]: I1204 05:23:08.300980 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxn5t\" (UniqueName: \"kubernetes.io/projected/5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2-kube-api-access-fxn5t\") pod \"crc-debug-657c4\" (UID: \"5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2\") " pod="openshift-must-gather-dz69k/crc-debug-657c4" Dec 04 05:23:08 crc kubenswrapper[4806]: I1204 05:23:08.382996 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dz69k/crc-debug-657c4" Dec 04 05:23:08 crc kubenswrapper[4806]: W1204 05:23:08.418575 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5de1fd92_d4a1_4e0a_bc4b_eeacbd995ff2.slice/crio-c83a7ad210a1d1a7c6d9a25e32e1bd43d101a8fb22f5a1ab47a1d65dfc8f0c16 WatchSource:0}: Error finding container c83a7ad210a1d1a7c6d9a25e32e1bd43d101a8fb22f5a1ab47a1d65dfc8f0c16: Status 404 returned error can't find the container with id c83a7ad210a1d1a7c6d9a25e32e1bd43d101a8fb22f5a1ab47a1d65dfc8f0c16 Dec 04 05:23:09 crc kubenswrapper[4806]: I1204 05:23:09.412523 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dz69k/crc-debug-657c4" event={"ID":"5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2","Type":"ContainerStarted","Data":"b126970236df12ff454a452033a8c9d0a2e3491303a7610137ef94c7c563cb7f"} Dec 04 05:23:09 crc kubenswrapper[4806]: I1204 05:23:09.413174 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dz69k/crc-debug-657c4" event={"ID":"5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2","Type":"ContainerStarted","Data":"c83a7ad210a1d1a7c6d9a25e32e1bd43d101a8fb22f5a1ab47a1d65dfc8f0c16"} Dec 04 05:23:49 crc kubenswrapper[4806]: I1204 05:23:49.781565 4806 generic.go:334] "Generic (PLEG): container finished" podID="5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2" containerID="b126970236df12ff454a452033a8c9d0a2e3491303a7610137ef94c7c563cb7f" exitCode=0 Dec 04 05:23:49 crc kubenswrapper[4806]: I1204 05:23:49.783005 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dz69k/crc-debug-657c4" event={"ID":"5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2","Type":"ContainerDied","Data":"b126970236df12ff454a452033a8c9d0a2e3491303a7610137ef94c7c563cb7f"} Dec 04 05:23:50 crc kubenswrapper[4806]: I1204 05:23:50.905455 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dz69k/crc-debug-657c4" Dec 04 05:23:50 crc kubenswrapper[4806]: I1204 05:23:50.924785 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxn5t\" (UniqueName: \"kubernetes.io/projected/5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2-kube-api-access-fxn5t\") pod \"5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2\" (UID: \"5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2\") " Dec 04 05:23:50 crc kubenswrapper[4806]: I1204 05:23:50.924872 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2-host\") pod \"5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2\" (UID: \"5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2\") " Dec 04 05:23:50 crc kubenswrapper[4806]: I1204 05:23:50.925213 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2-host" (OuterVolumeSpecName: "host") pod "5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2" (UID: "5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 05:23:50 crc kubenswrapper[4806]: I1204 05:23:50.930421 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2-kube-api-access-fxn5t" (OuterVolumeSpecName: "kube-api-access-fxn5t") pod "5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2" (UID: "5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2"). InnerVolumeSpecName "kube-api-access-fxn5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:23:50 crc kubenswrapper[4806]: I1204 05:23:50.955030 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-dz69k/crc-debug-657c4"] Dec 04 05:23:50 crc kubenswrapper[4806]: I1204 05:23:50.967632 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-dz69k/crc-debug-657c4"] Dec 04 05:23:51 crc kubenswrapper[4806]: I1204 05:23:51.026600 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxn5t\" (UniqueName: \"kubernetes.io/projected/5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2-kube-api-access-fxn5t\") on node \"crc\" DevicePath \"\"" Dec 04 05:23:51 crc kubenswrapper[4806]: I1204 05:23:51.026636 4806 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2-host\") on node \"crc\" DevicePath \"\"" Dec 04 05:23:51 crc kubenswrapper[4806]: I1204 05:23:51.432367 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2" path="/var/lib/kubelet/pods/5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2/volumes" Dec 04 05:23:51 crc kubenswrapper[4806]: I1204 05:23:51.813413 4806 scope.go:117] "RemoveContainer" containerID="b126970236df12ff454a452033a8c9d0a2e3491303a7610137ef94c7c563cb7f" Dec 04 05:23:51 crc kubenswrapper[4806]: I1204 05:23:51.813466 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dz69k/crc-debug-657c4" Dec 04 05:23:52 crc kubenswrapper[4806]: I1204 05:23:52.129430 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-dz69k/crc-debug-wlw7l"] Dec 04 05:23:52 crc kubenswrapper[4806]: E1204 05:23:52.129824 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2" containerName="container-00" Dec 04 05:23:52 crc kubenswrapper[4806]: I1204 05:23:52.129835 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2" containerName="container-00" Dec 04 05:23:52 crc kubenswrapper[4806]: I1204 05:23:52.130179 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="5de1fd92-d4a1-4e0a-bc4b-eeacbd995ff2" containerName="container-00" Dec 04 05:23:52 crc kubenswrapper[4806]: I1204 05:23:52.130746 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dz69k/crc-debug-wlw7l" Dec 04 05:23:52 crc kubenswrapper[4806]: I1204 05:23:52.132409 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-dz69k"/"default-dockercfg-fdnwj" Dec 04 05:23:52 crc kubenswrapper[4806]: I1204 05:23:52.244405 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq9g8\" (UniqueName: \"kubernetes.io/projected/cd409adb-81f7-464e-a911-8e900bdbe679-kube-api-access-qq9g8\") pod \"crc-debug-wlw7l\" (UID: \"cd409adb-81f7-464e-a911-8e900bdbe679\") " pod="openshift-must-gather-dz69k/crc-debug-wlw7l" Dec 04 05:23:52 crc kubenswrapper[4806]: I1204 05:23:52.244473 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd409adb-81f7-464e-a911-8e900bdbe679-host\") pod \"crc-debug-wlw7l\" (UID: \"cd409adb-81f7-464e-a911-8e900bdbe679\") " pod="openshift-must-gather-dz69k/crc-debug-wlw7l" Dec 04 05:23:52 crc kubenswrapper[4806]: I1204 05:23:52.346819 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq9g8\" (UniqueName: \"kubernetes.io/projected/cd409adb-81f7-464e-a911-8e900bdbe679-kube-api-access-qq9g8\") pod \"crc-debug-wlw7l\" (UID: \"cd409adb-81f7-464e-a911-8e900bdbe679\") " pod="openshift-must-gather-dz69k/crc-debug-wlw7l" Dec 04 05:23:52 crc kubenswrapper[4806]: I1204 05:23:52.346937 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd409adb-81f7-464e-a911-8e900bdbe679-host\") pod \"crc-debug-wlw7l\" (UID: \"cd409adb-81f7-464e-a911-8e900bdbe679\") " pod="openshift-must-gather-dz69k/crc-debug-wlw7l" Dec 04 05:23:52 crc kubenswrapper[4806]: I1204 05:23:52.347133 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd409adb-81f7-464e-a911-8e900bdbe679-host\") pod \"crc-debug-wlw7l\" (UID: \"cd409adb-81f7-464e-a911-8e900bdbe679\") " pod="openshift-must-gather-dz69k/crc-debug-wlw7l" Dec 04 05:23:52 crc kubenswrapper[4806]: I1204 05:23:52.366448 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq9g8\" (UniqueName: \"kubernetes.io/projected/cd409adb-81f7-464e-a911-8e900bdbe679-kube-api-access-qq9g8\") pod \"crc-debug-wlw7l\" (UID: \"cd409adb-81f7-464e-a911-8e900bdbe679\") " pod="openshift-must-gather-dz69k/crc-debug-wlw7l" Dec 04 05:23:52 crc kubenswrapper[4806]: I1204 05:23:52.447702 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dz69k/crc-debug-wlw7l" Dec 04 05:23:52 crc kubenswrapper[4806]: I1204 05:23:52.821507 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dz69k/crc-debug-wlw7l" event={"ID":"cd409adb-81f7-464e-a911-8e900bdbe679","Type":"ContainerStarted","Data":"c502d4a28e7e752800f8cd93930353745a4c61234a34ec1c5dba35e75a14a7ec"} Dec 04 05:23:53 crc kubenswrapper[4806]: I1204 05:23:53.834268 4806 generic.go:334] "Generic (PLEG): container finished" podID="cd409adb-81f7-464e-a911-8e900bdbe679" containerID="243d8cffc55529b9a7c58e34203ce3ab6b166fa34cf27c329467655cafd3762c" exitCode=0 Dec 04 05:23:53 crc kubenswrapper[4806]: I1204 05:23:53.834370 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dz69k/crc-debug-wlw7l" event={"ID":"cd409adb-81f7-464e-a911-8e900bdbe679","Type":"ContainerDied","Data":"243d8cffc55529b9a7c58e34203ce3ab6b166fa34cf27c329467655cafd3762c"} Dec 04 05:23:54 crc kubenswrapper[4806]: I1204 05:23:54.945988 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dz69k/crc-debug-wlw7l" Dec 04 05:23:55 crc kubenswrapper[4806]: I1204 05:23:55.087237 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qq9g8\" (UniqueName: \"kubernetes.io/projected/cd409adb-81f7-464e-a911-8e900bdbe679-kube-api-access-qq9g8\") pod \"cd409adb-81f7-464e-a911-8e900bdbe679\" (UID: \"cd409adb-81f7-464e-a911-8e900bdbe679\") " Dec 04 05:23:55 crc kubenswrapper[4806]: I1204 05:23:55.087602 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd409adb-81f7-464e-a911-8e900bdbe679-host\") pod \"cd409adb-81f7-464e-a911-8e900bdbe679\" (UID: \"cd409adb-81f7-464e-a911-8e900bdbe679\") " Dec 04 05:23:55 crc kubenswrapper[4806]: I1204 05:23:55.089744 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd409adb-81f7-464e-a911-8e900bdbe679-host" (OuterVolumeSpecName: "host") pod "cd409adb-81f7-464e-a911-8e900bdbe679" (UID: "cd409adb-81f7-464e-a911-8e900bdbe679"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 05:23:55 crc kubenswrapper[4806]: I1204 05:23:55.097695 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd409adb-81f7-464e-a911-8e900bdbe679-kube-api-access-qq9g8" (OuterVolumeSpecName: "kube-api-access-qq9g8") pod "cd409adb-81f7-464e-a911-8e900bdbe679" (UID: "cd409adb-81f7-464e-a911-8e900bdbe679"). InnerVolumeSpecName "kube-api-access-qq9g8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:23:55 crc kubenswrapper[4806]: I1204 05:23:55.191030 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qq9g8\" (UniqueName: \"kubernetes.io/projected/cd409adb-81f7-464e-a911-8e900bdbe679-kube-api-access-qq9g8\") on node \"crc\" DevicePath \"\"" Dec 04 05:23:55 crc kubenswrapper[4806]: I1204 05:23:55.191056 4806 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cd409adb-81f7-464e-a911-8e900bdbe679-host\") on node \"crc\" DevicePath \"\"" Dec 04 05:23:55 crc kubenswrapper[4806]: I1204 05:23:55.738996 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-dz69k/crc-debug-wlw7l"] Dec 04 05:23:55 crc kubenswrapper[4806]: I1204 05:23:55.748908 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-dz69k/crc-debug-wlw7l"] Dec 04 05:23:55 crc kubenswrapper[4806]: I1204 05:23:55.851149 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dz69k/crc-debug-wlw7l" Dec 04 05:23:55 crc kubenswrapper[4806]: I1204 05:23:55.851215 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c502d4a28e7e752800f8cd93930353745a4c61234a34ec1c5dba35e75a14a7ec" Dec 04 05:23:56 crc kubenswrapper[4806]: I1204 05:23:56.949435 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-dz69k/crc-debug-vccf2"] Dec 04 05:23:56 crc kubenswrapper[4806]: E1204 05:23:56.951356 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd409adb-81f7-464e-a911-8e900bdbe679" containerName="container-00" Dec 04 05:23:56 crc kubenswrapper[4806]: I1204 05:23:56.951376 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd409adb-81f7-464e-a911-8e900bdbe679" containerName="container-00" Dec 04 05:23:56 crc kubenswrapper[4806]: I1204 05:23:56.951590 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd409adb-81f7-464e-a911-8e900bdbe679" containerName="container-00" Dec 04 05:23:56 crc kubenswrapper[4806]: I1204 05:23:56.952173 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dz69k/crc-debug-vccf2" Dec 04 05:23:56 crc kubenswrapper[4806]: I1204 05:23:56.954467 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-dz69k"/"default-dockercfg-fdnwj" Dec 04 05:23:57 crc kubenswrapper[4806]: I1204 05:23:57.124449 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz5fr\" (UniqueName: \"kubernetes.io/projected/2ef9c4d0-0cc2-49e6-b320-49d454624999-kube-api-access-sz5fr\") pod \"crc-debug-vccf2\" (UID: \"2ef9c4d0-0cc2-49e6-b320-49d454624999\") " pod="openshift-must-gather-dz69k/crc-debug-vccf2" Dec 04 05:23:57 crc kubenswrapper[4806]: I1204 05:23:57.124528 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2ef9c4d0-0cc2-49e6-b320-49d454624999-host\") pod \"crc-debug-vccf2\" (UID: \"2ef9c4d0-0cc2-49e6-b320-49d454624999\") " pod="openshift-must-gather-dz69k/crc-debug-vccf2" Dec 04 05:23:57 crc kubenswrapper[4806]: I1204 05:23:57.225789 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2ef9c4d0-0cc2-49e6-b320-49d454624999-host\") pod \"crc-debug-vccf2\" (UID: \"2ef9c4d0-0cc2-49e6-b320-49d454624999\") " pod="openshift-must-gather-dz69k/crc-debug-vccf2" Dec 04 05:23:57 crc kubenswrapper[4806]: I1204 05:23:57.225917 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2ef9c4d0-0cc2-49e6-b320-49d454624999-host\") pod \"crc-debug-vccf2\" (UID: \"2ef9c4d0-0cc2-49e6-b320-49d454624999\") " pod="openshift-must-gather-dz69k/crc-debug-vccf2" Dec 04 05:23:57 crc kubenswrapper[4806]: I1204 05:23:57.225996 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz5fr\" (UniqueName: \"kubernetes.io/projected/2ef9c4d0-0cc2-49e6-b320-49d454624999-kube-api-access-sz5fr\") pod \"crc-debug-vccf2\" (UID: \"2ef9c4d0-0cc2-49e6-b320-49d454624999\") " pod="openshift-must-gather-dz69k/crc-debug-vccf2" Dec 04 05:23:57 crc kubenswrapper[4806]: I1204 05:23:57.244617 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz5fr\" (UniqueName: \"kubernetes.io/projected/2ef9c4d0-0cc2-49e6-b320-49d454624999-kube-api-access-sz5fr\") pod \"crc-debug-vccf2\" (UID: \"2ef9c4d0-0cc2-49e6-b320-49d454624999\") " pod="openshift-must-gather-dz69k/crc-debug-vccf2" Dec 04 05:23:57 crc kubenswrapper[4806]: I1204 05:23:57.271336 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dz69k/crc-debug-vccf2" Dec 04 05:23:57 crc kubenswrapper[4806]: I1204 05:23:57.457117 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd409adb-81f7-464e-a911-8e900bdbe679" path="/var/lib/kubelet/pods/cd409adb-81f7-464e-a911-8e900bdbe679/volumes" Dec 04 05:23:57 crc kubenswrapper[4806]: I1204 05:23:57.869109 4806 generic.go:334] "Generic (PLEG): container finished" podID="2ef9c4d0-0cc2-49e6-b320-49d454624999" containerID="f9ebec8313cbe7eba40e628f6666453ccb40e95fdc1386f31d1bb61e2970a70f" exitCode=0 Dec 04 05:23:57 crc kubenswrapper[4806]: I1204 05:23:57.869192 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dz69k/crc-debug-vccf2" event={"ID":"2ef9c4d0-0cc2-49e6-b320-49d454624999","Type":"ContainerDied","Data":"f9ebec8313cbe7eba40e628f6666453ccb40e95fdc1386f31d1bb61e2970a70f"} Dec 04 05:23:57 crc kubenswrapper[4806]: I1204 05:23:57.869466 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dz69k/crc-debug-vccf2" event={"ID":"2ef9c4d0-0cc2-49e6-b320-49d454624999","Type":"ContainerStarted","Data":"1fc819a27a170d3ad4f1d81067971dfeffdd7515c275d87029cebf807e6f92b0"} Dec 04 05:23:57 crc kubenswrapper[4806]: I1204 05:23:57.912502 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-dz69k/crc-debug-vccf2"] Dec 04 05:23:57 crc kubenswrapper[4806]: I1204 05:23:57.922494 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-dz69k/crc-debug-vccf2"] Dec 04 05:23:58 crc kubenswrapper[4806]: I1204 05:23:58.998442 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dz69k/crc-debug-vccf2" Dec 04 05:23:59 crc kubenswrapper[4806]: I1204 05:23:59.164398 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sz5fr\" (UniqueName: \"kubernetes.io/projected/2ef9c4d0-0cc2-49e6-b320-49d454624999-kube-api-access-sz5fr\") pod \"2ef9c4d0-0cc2-49e6-b320-49d454624999\" (UID: \"2ef9c4d0-0cc2-49e6-b320-49d454624999\") " Dec 04 05:23:59 crc kubenswrapper[4806]: I1204 05:23:59.164432 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2ef9c4d0-0cc2-49e6-b320-49d454624999-host\") pod \"2ef9c4d0-0cc2-49e6-b320-49d454624999\" (UID: \"2ef9c4d0-0cc2-49e6-b320-49d454624999\") " Dec 04 05:23:59 crc kubenswrapper[4806]: I1204 05:23:59.165158 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2ef9c4d0-0cc2-49e6-b320-49d454624999-host" (OuterVolumeSpecName: "host") pod "2ef9c4d0-0cc2-49e6-b320-49d454624999" (UID: "2ef9c4d0-0cc2-49e6-b320-49d454624999"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 05:23:59 crc kubenswrapper[4806]: I1204 05:23:59.205807 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ef9c4d0-0cc2-49e6-b320-49d454624999-kube-api-access-sz5fr" (OuterVolumeSpecName: "kube-api-access-sz5fr") pod "2ef9c4d0-0cc2-49e6-b320-49d454624999" (UID: "2ef9c4d0-0cc2-49e6-b320-49d454624999"). InnerVolumeSpecName "kube-api-access-sz5fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:23:59 crc kubenswrapper[4806]: I1204 05:23:59.267414 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sz5fr\" (UniqueName: \"kubernetes.io/projected/2ef9c4d0-0cc2-49e6-b320-49d454624999-kube-api-access-sz5fr\") on node \"crc\" DevicePath \"\"" Dec 04 05:23:59 crc kubenswrapper[4806]: I1204 05:23:59.267758 4806 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2ef9c4d0-0cc2-49e6-b320-49d454624999-host\") on node \"crc\" DevicePath \"\"" Dec 04 05:23:59 crc kubenswrapper[4806]: I1204 05:23:59.434223 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ef9c4d0-0cc2-49e6-b320-49d454624999" path="/var/lib/kubelet/pods/2ef9c4d0-0cc2-49e6-b320-49d454624999/volumes" Dec 04 05:23:59 crc kubenswrapper[4806]: I1204 05:23:59.884737 4806 scope.go:117] "RemoveContainer" containerID="f9ebec8313cbe7eba40e628f6666453ccb40e95fdc1386f31d1bb61e2970a70f" Dec 04 05:23:59 crc kubenswrapper[4806]: I1204 05:23:59.884770 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dz69k/crc-debug-vccf2" Dec 04 05:24:00 crc kubenswrapper[4806]: I1204 05:24:00.077800 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sq9hc"] Dec 04 05:24:00 crc kubenswrapper[4806]: E1204 05:24:00.078248 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ef9c4d0-0cc2-49e6-b320-49d454624999" containerName="container-00" Dec 04 05:24:00 crc kubenswrapper[4806]: I1204 05:24:00.078265 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ef9c4d0-0cc2-49e6-b320-49d454624999" containerName="container-00" Dec 04 05:24:00 crc kubenswrapper[4806]: I1204 05:24:00.078565 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ef9c4d0-0cc2-49e6-b320-49d454624999" containerName="container-00" Dec 04 05:24:00 crc kubenswrapper[4806]: I1204 05:24:00.081742 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:00 crc kubenswrapper[4806]: I1204 05:24:00.089689 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sq9hc"] Dec 04 05:24:00 crc kubenswrapper[4806]: I1204 05:24:00.185093 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s94r9\" (UniqueName: \"kubernetes.io/projected/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-kube-api-access-s94r9\") pod \"redhat-operators-sq9hc\" (UID: \"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec\") " pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:00 crc kubenswrapper[4806]: I1204 05:24:00.185563 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-utilities\") pod \"redhat-operators-sq9hc\" (UID: \"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec\") " pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:00 crc kubenswrapper[4806]: I1204 05:24:00.185682 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-catalog-content\") pod \"redhat-operators-sq9hc\" (UID: \"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec\") " pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:00 crc kubenswrapper[4806]: I1204 05:24:00.287083 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-catalog-content\") pod \"redhat-operators-sq9hc\" (UID: \"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec\") " pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:00 crc kubenswrapper[4806]: I1204 05:24:00.287176 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s94r9\" (UniqueName: \"kubernetes.io/projected/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-kube-api-access-s94r9\") pod \"redhat-operators-sq9hc\" (UID: \"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec\") " pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:00 crc kubenswrapper[4806]: I1204 05:24:00.287298 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-utilities\") pod \"redhat-operators-sq9hc\" (UID: \"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec\") " pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:00 crc kubenswrapper[4806]: I1204 05:24:00.287489 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-catalog-content\") pod \"redhat-operators-sq9hc\" (UID: \"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec\") " pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:00 crc kubenswrapper[4806]: I1204 05:24:00.287698 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-utilities\") pod \"redhat-operators-sq9hc\" (UID: \"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec\") " pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:00 crc kubenswrapper[4806]: I1204 05:24:00.305766 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s94r9\" (UniqueName: \"kubernetes.io/projected/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-kube-api-access-s94r9\") pod \"redhat-operators-sq9hc\" (UID: \"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec\") " pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:00 crc kubenswrapper[4806]: I1204 05:24:00.398975 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:00 crc kubenswrapper[4806]: I1204 05:24:00.924721 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sq9hc"] Dec 04 05:24:00 crc kubenswrapper[4806]: W1204 05:24:00.939306 4806 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod912c77e8_c9e7_49dd_aa2c_d5745f10b2ec.slice/crio-cda1cd4d0647093e4bc25ead65c73cc8add94fb11e577354c428488950a2029f WatchSource:0}: Error finding container cda1cd4d0647093e4bc25ead65c73cc8add94fb11e577354c428488950a2029f: Status 404 returned error can't find the container with id cda1cd4d0647093e4bc25ead65c73cc8add94fb11e577354c428488950a2029f Dec 04 05:24:01 crc kubenswrapper[4806]: I1204 05:24:01.902893 4806 generic.go:334] "Generic (PLEG): container finished" podID="912c77e8-c9e7-49dd-aa2c-d5745f10b2ec" containerID="b9495cac74ba25935f01c703ce468b641514e3de532db6a405c3fdb56003f73b" exitCode=0 Dec 04 05:24:01 crc kubenswrapper[4806]: I1204 05:24:01.902962 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sq9hc" event={"ID":"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec","Type":"ContainerDied","Data":"b9495cac74ba25935f01c703ce468b641514e3de532db6a405c3fdb56003f73b"} Dec 04 05:24:01 crc kubenswrapper[4806]: I1204 05:24:01.903044 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sq9hc" event={"ID":"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec","Type":"ContainerStarted","Data":"cda1cd4d0647093e4bc25ead65c73cc8add94fb11e577354c428488950a2029f"} Dec 04 05:24:01 crc kubenswrapper[4806]: I1204 05:24:01.904871 4806 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 05:24:02 crc kubenswrapper[4806]: I1204 05:24:02.914760 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sq9hc" event={"ID":"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec","Type":"ContainerStarted","Data":"7bf901288a6e6e7fab1270d5e17fcb07061950dba10e57dd530a3d61899dc880"} Dec 04 05:24:05 crc kubenswrapper[4806]: I1204 05:24:05.938965 4806 generic.go:334] "Generic (PLEG): container finished" podID="912c77e8-c9e7-49dd-aa2c-d5745f10b2ec" containerID="7bf901288a6e6e7fab1270d5e17fcb07061950dba10e57dd530a3d61899dc880" exitCode=0 Dec 04 05:24:05 crc kubenswrapper[4806]: I1204 05:24:05.939468 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sq9hc" event={"ID":"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec","Type":"ContainerDied","Data":"7bf901288a6e6e7fab1270d5e17fcb07061950dba10e57dd530a3d61899dc880"} Dec 04 05:24:06 crc kubenswrapper[4806]: I1204 05:24:06.955373 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sq9hc" event={"ID":"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec","Type":"ContainerStarted","Data":"360548fb6a0a24eb06a39a0484dad88d6abeb7e98085e5e68b10ef772b326ab8"} Dec 04 05:24:06 crc kubenswrapper[4806]: I1204 05:24:06.973813 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sq9hc" podStartSLOduration=2.448018042 podStartE2EDuration="6.973792001s" podCreationTimestamp="2025-12-04 05:24:00 +0000 UTC" firstStartedPulling="2025-12-04 05:24:01.904666874 +0000 UTC m=+5356.763179802" lastFinishedPulling="2025-12-04 05:24:06.430440833 +0000 UTC m=+5361.288953761" observedRunningTime="2025-12-04 05:24:06.971356144 +0000 UTC m=+5361.829869072" watchObservedRunningTime="2025-12-04 05:24:06.973792001 +0000 UTC m=+5361.832304929" Dec 04 05:24:10 crc kubenswrapper[4806]: I1204 05:24:10.399330 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:10 crc kubenswrapper[4806]: I1204 05:24:10.399850 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:11 crc kubenswrapper[4806]: I1204 05:24:11.466314 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sq9hc" podUID="912c77e8-c9e7-49dd-aa2c-d5745f10b2ec" containerName="registry-server" probeResult="failure" output=< Dec 04 05:24:11 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Dec 04 05:24:11 crc kubenswrapper[4806]: > Dec 04 05:24:20 crc kubenswrapper[4806]: I1204 05:24:20.448511 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:20 crc kubenswrapper[4806]: I1204 05:24:20.504119 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:20 crc kubenswrapper[4806]: I1204 05:24:20.685733 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sq9hc"] Dec 04 05:24:22 crc kubenswrapper[4806]: I1204 05:24:22.085140 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sq9hc" podUID="912c77e8-c9e7-49dd-aa2c-d5745f10b2ec" containerName="registry-server" containerID="cri-o://360548fb6a0a24eb06a39a0484dad88d6abeb7e98085e5e68b10ef772b326ab8" gracePeriod=2 Dec 04 05:24:22 crc kubenswrapper[4806]: I1204 05:24:22.557304 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:22 crc kubenswrapper[4806]: I1204 05:24:22.738072 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-catalog-content\") pod \"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec\" (UID: \"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec\") " Dec 04 05:24:22 crc kubenswrapper[4806]: I1204 05:24:22.738426 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s94r9\" (UniqueName: \"kubernetes.io/projected/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-kube-api-access-s94r9\") pod \"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec\" (UID: \"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec\") " Dec 04 05:24:22 crc kubenswrapper[4806]: I1204 05:24:22.738504 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-utilities\") pod \"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec\" (UID: \"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec\") " Dec 04 05:24:22 crc kubenswrapper[4806]: I1204 05:24:22.739414 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-utilities" (OuterVolumeSpecName: "utilities") pod "912c77e8-c9e7-49dd-aa2c-d5745f10b2ec" (UID: "912c77e8-c9e7-49dd-aa2c-d5745f10b2ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:24:22 crc kubenswrapper[4806]: I1204 05:24:22.748782 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-kube-api-access-s94r9" (OuterVolumeSpecName: "kube-api-access-s94r9") pod "912c77e8-c9e7-49dd-aa2c-d5745f10b2ec" (UID: "912c77e8-c9e7-49dd-aa2c-d5745f10b2ec"). InnerVolumeSpecName "kube-api-access-s94r9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:24:22 crc kubenswrapper[4806]: I1204 05:24:22.841025 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s94r9\" (UniqueName: \"kubernetes.io/projected/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-kube-api-access-s94r9\") on node \"crc\" DevicePath \"\"" Dec 04 05:24:22 crc kubenswrapper[4806]: I1204 05:24:22.841309 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 05:24:22 crc kubenswrapper[4806]: I1204 05:24:22.880115 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "912c77e8-c9e7-49dd-aa2c-d5745f10b2ec" (UID: "912c77e8-c9e7-49dd-aa2c-d5745f10b2ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:24:22 crc kubenswrapper[4806]: I1204 05:24:22.944233 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 05:24:23 crc kubenswrapper[4806]: I1204 05:24:23.112706 4806 generic.go:334] "Generic (PLEG): container finished" podID="912c77e8-c9e7-49dd-aa2c-d5745f10b2ec" containerID="360548fb6a0a24eb06a39a0484dad88d6abeb7e98085e5e68b10ef772b326ab8" exitCode=0 Dec 04 05:24:23 crc kubenswrapper[4806]: I1204 05:24:23.112780 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sq9hc" event={"ID":"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec","Type":"ContainerDied","Data":"360548fb6a0a24eb06a39a0484dad88d6abeb7e98085e5e68b10ef772b326ab8"} Dec 04 05:24:23 crc kubenswrapper[4806]: I1204 05:24:23.112859 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sq9hc" Dec 04 05:24:23 crc kubenswrapper[4806]: I1204 05:24:23.112886 4806 scope.go:117] "RemoveContainer" containerID="360548fb6a0a24eb06a39a0484dad88d6abeb7e98085e5e68b10ef772b326ab8" Dec 04 05:24:23 crc kubenswrapper[4806]: I1204 05:24:23.112867 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sq9hc" event={"ID":"912c77e8-c9e7-49dd-aa2c-d5745f10b2ec","Type":"ContainerDied","Data":"cda1cd4d0647093e4bc25ead65c73cc8add94fb11e577354c428488950a2029f"} Dec 04 05:24:23 crc kubenswrapper[4806]: I1204 05:24:23.145961 4806 scope.go:117] "RemoveContainer" containerID="7bf901288a6e6e7fab1270d5e17fcb07061950dba10e57dd530a3d61899dc880" Dec 04 05:24:23 crc kubenswrapper[4806]: I1204 05:24:23.175116 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sq9hc"] Dec 04 05:24:23 crc kubenswrapper[4806]: I1204 05:24:23.183268 4806 scope.go:117] "RemoveContainer" containerID="b9495cac74ba25935f01c703ce468b641514e3de532db6a405c3fdb56003f73b" Dec 04 05:24:23 crc kubenswrapper[4806]: I1204 05:24:23.194824 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sq9hc"] Dec 04 05:24:23 crc kubenswrapper[4806]: I1204 05:24:23.213737 4806 scope.go:117] "RemoveContainer" containerID="360548fb6a0a24eb06a39a0484dad88d6abeb7e98085e5e68b10ef772b326ab8" Dec 04 05:24:23 crc kubenswrapper[4806]: E1204 05:24:23.214389 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"360548fb6a0a24eb06a39a0484dad88d6abeb7e98085e5e68b10ef772b326ab8\": container with ID starting with 360548fb6a0a24eb06a39a0484dad88d6abeb7e98085e5e68b10ef772b326ab8 not found: ID does not exist" containerID="360548fb6a0a24eb06a39a0484dad88d6abeb7e98085e5e68b10ef772b326ab8" Dec 04 05:24:23 crc kubenswrapper[4806]: I1204 05:24:23.214444 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"360548fb6a0a24eb06a39a0484dad88d6abeb7e98085e5e68b10ef772b326ab8"} err="failed to get container status \"360548fb6a0a24eb06a39a0484dad88d6abeb7e98085e5e68b10ef772b326ab8\": rpc error: code = NotFound desc = could not find container \"360548fb6a0a24eb06a39a0484dad88d6abeb7e98085e5e68b10ef772b326ab8\": container with ID starting with 360548fb6a0a24eb06a39a0484dad88d6abeb7e98085e5e68b10ef772b326ab8 not found: ID does not exist" Dec 04 05:24:23 crc kubenswrapper[4806]: I1204 05:24:23.214473 4806 scope.go:117] "RemoveContainer" containerID="7bf901288a6e6e7fab1270d5e17fcb07061950dba10e57dd530a3d61899dc880" Dec 04 05:24:23 crc kubenswrapper[4806]: E1204 05:24:23.214800 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bf901288a6e6e7fab1270d5e17fcb07061950dba10e57dd530a3d61899dc880\": container with ID starting with 7bf901288a6e6e7fab1270d5e17fcb07061950dba10e57dd530a3d61899dc880 not found: ID does not exist" containerID="7bf901288a6e6e7fab1270d5e17fcb07061950dba10e57dd530a3d61899dc880" Dec 04 05:24:23 crc kubenswrapper[4806]: I1204 05:24:23.214836 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bf901288a6e6e7fab1270d5e17fcb07061950dba10e57dd530a3d61899dc880"} err="failed to get container status \"7bf901288a6e6e7fab1270d5e17fcb07061950dba10e57dd530a3d61899dc880\": rpc error: code = NotFound desc = could not find container \"7bf901288a6e6e7fab1270d5e17fcb07061950dba10e57dd530a3d61899dc880\": container with ID starting with 7bf901288a6e6e7fab1270d5e17fcb07061950dba10e57dd530a3d61899dc880 not found: ID does not exist" Dec 04 05:24:23 crc kubenswrapper[4806]: I1204 05:24:23.214859 4806 scope.go:117] "RemoveContainer" containerID="b9495cac74ba25935f01c703ce468b641514e3de532db6a405c3fdb56003f73b" Dec 04 05:24:23 crc kubenswrapper[4806]: E1204 05:24:23.215163 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9495cac74ba25935f01c703ce468b641514e3de532db6a405c3fdb56003f73b\": container with ID starting with b9495cac74ba25935f01c703ce468b641514e3de532db6a405c3fdb56003f73b not found: ID does not exist" containerID="b9495cac74ba25935f01c703ce468b641514e3de532db6a405c3fdb56003f73b" Dec 04 05:24:23 crc kubenswrapper[4806]: I1204 05:24:23.215182 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9495cac74ba25935f01c703ce468b641514e3de532db6a405c3fdb56003f73b"} err="failed to get container status \"b9495cac74ba25935f01c703ce468b641514e3de532db6a405c3fdb56003f73b\": rpc error: code = NotFound desc = could not find container \"b9495cac74ba25935f01c703ce468b641514e3de532db6a405c3fdb56003f73b\": container with ID starting with b9495cac74ba25935f01c703ce468b641514e3de532db6a405c3fdb56003f73b not found: ID does not exist" Dec 04 05:24:23 crc kubenswrapper[4806]: I1204 05:24:23.436242 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="912c77e8-c9e7-49dd-aa2c-d5745f10b2ec" path="/var/lib/kubelet/pods/912c77e8-c9e7-49dd-aa2c-d5745f10b2ec/volumes" Dec 04 05:24:27 crc kubenswrapper[4806]: I1204 05:24:27.046987 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:24:27 crc kubenswrapper[4806]: I1204 05:24:27.048503 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.395331 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2zzsv"] Dec 04 05:24:36 crc kubenswrapper[4806]: E1204 05:24:36.398096 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="912c77e8-c9e7-49dd-aa2c-d5745f10b2ec" containerName="registry-server" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.398266 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="912c77e8-c9e7-49dd-aa2c-d5745f10b2ec" containerName="registry-server" Dec 04 05:24:36 crc kubenswrapper[4806]: E1204 05:24:36.398413 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="912c77e8-c9e7-49dd-aa2c-d5745f10b2ec" containerName="extract-content" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.398574 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="912c77e8-c9e7-49dd-aa2c-d5745f10b2ec" containerName="extract-content" Dec 04 05:24:36 crc kubenswrapper[4806]: E1204 05:24:36.398744 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="912c77e8-c9e7-49dd-aa2c-d5745f10b2ec" containerName="extract-utilities" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.398869 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="912c77e8-c9e7-49dd-aa2c-d5745f10b2ec" containerName="extract-utilities" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.399390 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="912c77e8-c9e7-49dd-aa2c-d5745f10b2ec" containerName="registry-server" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.402011 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.420841 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2zzsv"] Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.441849 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhzxf\" (UniqueName: \"kubernetes.io/projected/3c5e571a-32a2-401e-9370-af762975718f-kube-api-access-nhzxf\") pod \"certified-operators-2zzsv\" (UID: \"3c5e571a-32a2-401e-9370-af762975718f\") " pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.441991 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5e571a-32a2-401e-9370-af762975718f-utilities\") pod \"certified-operators-2zzsv\" (UID: \"3c5e571a-32a2-401e-9370-af762975718f\") " pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.442065 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5e571a-32a2-401e-9370-af762975718f-catalog-content\") pod \"certified-operators-2zzsv\" (UID: \"3c5e571a-32a2-401e-9370-af762975718f\") " pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.544255 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhzxf\" (UniqueName: \"kubernetes.io/projected/3c5e571a-32a2-401e-9370-af762975718f-kube-api-access-nhzxf\") pod \"certified-operators-2zzsv\" (UID: \"3c5e571a-32a2-401e-9370-af762975718f\") " pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.544354 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5e571a-32a2-401e-9370-af762975718f-utilities\") pod \"certified-operators-2zzsv\" (UID: \"3c5e571a-32a2-401e-9370-af762975718f\") " pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.544469 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5e571a-32a2-401e-9370-af762975718f-catalog-content\") pod \"certified-operators-2zzsv\" (UID: \"3c5e571a-32a2-401e-9370-af762975718f\") " pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.544757 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5e571a-32a2-401e-9370-af762975718f-utilities\") pod \"certified-operators-2zzsv\" (UID: \"3c5e571a-32a2-401e-9370-af762975718f\") " pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.545017 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5e571a-32a2-401e-9370-af762975718f-catalog-content\") pod \"certified-operators-2zzsv\" (UID: \"3c5e571a-32a2-401e-9370-af762975718f\") " pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.570161 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hf8px"] Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.576357 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.581492 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhzxf\" (UniqueName: \"kubernetes.io/projected/3c5e571a-32a2-401e-9370-af762975718f-kube-api-access-nhzxf\") pod \"certified-operators-2zzsv\" (UID: \"3c5e571a-32a2-401e-9370-af762975718f\") " pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.591058 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hf8px"] Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.649168 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c044b665-2475-438c-9616-f8856947509a-utilities\") pod \"community-operators-hf8px\" (UID: \"c044b665-2475-438c-9616-f8856947509a\") " pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.649242 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c044b665-2475-438c-9616-f8856947509a-catalog-content\") pod \"community-operators-hf8px\" (UID: \"c044b665-2475-438c-9616-f8856947509a\") " pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.649299 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmrxk\" (UniqueName: \"kubernetes.io/projected/c044b665-2475-438c-9616-f8856947509a-kube-api-access-zmrxk\") pod \"community-operators-hf8px\" (UID: \"c044b665-2475-438c-9616-f8856947509a\") " pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.734816 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.750752 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c044b665-2475-438c-9616-f8856947509a-utilities\") pod \"community-operators-hf8px\" (UID: \"c044b665-2475-438c-9616-f8856947509a\") " pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.750797 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c044b665-2475-438c-9616-f8856947509a-catalog-content\") pod \"community-operators-hf8px\" (UID: \"c044b665-2475-438c-9616-f8856947509a\") " pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.750837 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmrxk\" (UniqueName: \"kubernetes.io/projected/c044b665-2475-438c-9616-f8856947509a-kube-api-access-zmrxk\") pod \"community-operators-hf8px\" (UID: \"c044b665-2475-438c-9616-f8856947509a\") " pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.751739 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c044b665-2475-438c-9616-f8856947509a-utilities\") pod \"community-operators-hf8px\" (UID: \"c044b665-2475-438c-9616-f8856947509a\") " pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.751963 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c044b665-2475-438c-9616-f8856947509a-catalog-content\") pod \"community-operators-hf8px\" (UID: \"c044b665-2475-438c-9616-f8856947509a\") " pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.766434 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmrxk\" (UniqueName: \"kubernetes.io/projected/c044b665-2475-438c-9616-f8856947509a-kube-api-access-zmrxk\") pod \"community-operators-hf8px\" (UID: \"c044b665-2475-438c-9616-f8856947509a\") " pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:24:36 crc kubenswrapper[4806]: I1204 05:24:36.960458 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:24:37 crc kubenswrapper[4806]: I1204 05:24:37.382348 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2zzsv"] Dec 04 05:24:37 crc kubenswrapper[4806]: I1204 05:24:37.661197 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hf8px"] Dec 04 05:24:38 crc kubenswrapper[4806]: I1204 05:24:38.249098 4806 generic.go:334] "Generic (PLEG): container finished" podID="3c5e571a-32a2-401e-9370-af762975718f" containerID="5eb53285a1df641afad87d6d04ea0419da44b734d5f8be63547c6dae6a91faef" exitCode=0 Dec 04 05:24:38 crc kubenswrapper[4806]: I1204 05:24:38.249165 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2zzsv" event={"ID":"3c5e571a-32a2-401e-9370-af762975718f","Type":"ContainerDied","Data":"5eb53285a1df641afad87d6d04ea0419da44b734d5f8be63547c6dae6a91faef"} Dec 04 05:24:38 crc kubenswrapper[4806]: I1204 05:24:38.249419 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2zzsv" event={"ID":"3c5e571a-32a2-401e-9370-af762975718f","Type":"ContainerStarted","Data":"291b71357975a2e0e8c3231728b1853dc5888f16fabfd14adefa98a6299d57cb"} Dec 04 05:24:38 crc kubenswrapper[4806]: I1204 05:24:38.252127 4806 generic.go:334] "Generic (PLEG): container finished" podID="c044b665-2475-438c-9616-f8856947509a" containerID="bdf9a16d6d82d2142ab9ac9771f5a70211a71c8d71df85a13e985b680438d26c" exitCode=0 Dec 04 05:24:38 crc kubenswrapper[4806]: I1204 05:24:38.252154 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf8px" event={"ID":"c044b665-2475-438c-9616-f8856947509a","Type":"ContainerDied","Data":"bdf9a16d6d82d2142ab9ac9771f5a70211a71c8d71df85a13e985b680438d26c"} Dec 04 05:24:38 crc kubenswrapper[4806]: I1204 05:24:38.252172 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf8px" event={"ID":"c044b665-2475-438c-9616-f8856947509a","Type":"ContainerStarted","Data":"c119a31ae2c73abb1509e62666e49f31bf7b91d621dae22c95abe5b77d8ce246"} Dec 04 05:24:39 crc kubenswrapper[4806]: I1204 05:24:39.260895 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2zzsv" event={"ID":"3c5e571a-32a2-401e-9370-af762975718f","Type":"ContainerStarted","Data":"f6441f48494b257e10f5db8abc4b6d31e7a028de70781b9b54b7331146dd87ce"} Dec 04 05:24:39 crc kubenswrapper[4806]: I1204 05:24:39.264131 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf8px" event={"ID":"c044b665-2475-438c-9616-f8856947509a","Type":"ContainerStarted","Data":"b82e25d3ddcad7fb5643bfe3aebde7e140f5b309bde0b3a5a0a06232538d317f"} Dec 04 05:24:40 crc kubenswrapper[4806]: I1204 05:24:40.445992 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-949b6795d-z7dx2_ef61aa0b-2b2c-481f-b22d-4ea770a91711/barbican-api/0.log" Dec 04 05:24:40 crc kubenswrapper[4806]: I1204 05:24:40.567517 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-949b6795d-z7dx2_ef61aa0b-2b2c-481f-b22d-4ea770a91711/barbican-api-log/0.log" Dec 04 05:24:40 crc kubenswrapper[4806]: I1204 05:24:40.679709 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-b658d66b-ft4r8_fef48eb8-ea9d-42ab-8839-fa7a890fe301/barbican-keystone-listener/0.log" Dec 04 05:24:40 crc kubenswrapper[4806]: I1204 05:24:40.804334 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-b658d66b-ft4r8_fef48eb8-ea9d-42ab-8839-fa7a890fe301/barbican-keystone-listener-log/0.log" Dec 04 05:24:40 crc kubenswrapper[4806]: I1204 05:24:40.872670 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6785cc669c-hgz8x_cc467d3b-daf0-4869-9cc3-1034cdfd373a/barbican-worker/0.log" Dec 04 05:24:40 crc kubenswrapper[4806]: I1204 05:24:40.908587 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6785cc669c-hgz8x_cc467d3b-daf0-4869-9cc3-1034cdfd373a/barbican-worker-log/0.log" Dec 04 05:24:41 crc kubenswrapper[4806]: I1204 05:24:41.326044 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_56b1c86b-9508-4ca9-a2b8-9a496875c9e8/ceilometer-central-agent/0.log" Dec 04 05:24:41 crc kubenswrapper[4806]: I1204 05:24:41.420166 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-lbsn8_1787b37a-3067-4912-a658-4a8fd53faab6/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:24:41 crc kubenswrapper[4806]: I1204 05:24:41.425209 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_56b1c86b-9508-4ca9-a2b8-9a496875c9e8/proxy-httpd/0.log" Dec 04 05:24:41 crc kubenswrapper[4806]: I1204 05:24:41.448002 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_56b1c86b-9508-4ca9-a2b8-9a496875c9e8/ceilometer-notification-agent/0.log" Dec 04 05:24:41 crc kubenswrapper[4806]: I1204 05:24:41.509251 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_56b1c86b-9508-4ca9-a2b8-9a496875c9e8/sg-core/0.log" Dec 04 05:24:41 crc kubenswrapper[4806]: I1204 05:24:41.760104 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3d05a5b8-2294-4d30-a0b7-d865c8aae646/cinder-api-log/0.log" Dec 04 05:24:41 crc kubenswrapper[4806]: I1204 05:24:41.797039 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3d05a5b8-2294-4d30-a0b7-d865c8aae646/cinder-api/0.log" Dec 04 05:24:41 crc kubenswrapper[4806]: I1204 05:24:41.967392 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8205f150-c0db-445a-b005-30d0632610c0/cinder-scheduler/0.log" Dec 04 05:24:42 crc kubenswrapper[4806]: I1204 05:24:42.038858 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8205f150-c0db-445a-b005-30d0632610c0/probe/0.log" Dec 04 05:24:42 crc kubenswrapper[4806]: I1204 05:24:42.105152 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-vcgs6_bbebb80f-3f15-464c-bb52-ab7dc2a100c4/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:24:42 crc kubenswrapper[4806]: I1204 05:24:42.261735 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-984f8_cc3788a8-d331-4c5c-a80b-d73e1ef5da9c/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:24:42 crc kubenswrapper[4806]: I1204 05:24:42.310672 4806 generic.go:334] "Generic (PLEG): container finished" podID="3c5e571a-32a2-401e-9370-af762975718f" containerID="f6441f48494b257e10f5db8abc4b6d31e7a028de70781b9b54b7331146dd87ce" exitCode=0 Dec 04 05:24:42 crc kubenswrapper[4806]: I1204 05:24:42.310748 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2zzsv" event={"ID":"3c5e571a-32a2-401e-9370-af762975718f","Type":"ContainerDied","Data":"f6441f48494b257e10f5db8abc4b6d31e7a028de70781b9b54b7331146dd87ce"} Dec 04 05:24:42 crc kubenswrapper[4806]: I1204 05:24:42.313000 4806 generic.go:334] "Generic (PLEG): container finished" podID="c044b665-2475-438c-9616-f8856947509a" containerID="b82e25d3ddcad7fb5643bfe3aebde7e140f5b309bde0b3a5a0a06232538d317f" exitCode=0 Dec 04 05:24:42 crc kubenswrapper[4806]: I1204 05:24:42.313031 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf8px" event={"ID":"c044b665-2475-438c-9616-f8856947509a","Type":"ContainerDied","Data":"b82e25d3ddcad7fb5643bfe3aebde7e140f5b309bde0b3a5a0a06232538d317f"} Dec 04 05:24:42 crc kubenswrapper[4806]: I1204 05:24:42.368565 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7677974f-lk2tx_5e42fed2-6410-480b-ad51-7c8a5c383eea/init/0.log" Dec 04 05:24:42 crc kubenswrapper[4806]: I1204 05:24:42.609102 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7677974f-lk2tx_5e42fed2-6410-480b-ad51-7c8a5c383eea/init/0.log" Dec 04 05:24:42 crc kubenswrapper[4806]: I1204 05:24:42.702896 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-pv9wv_0b8e6232-8e76-425c-9b02-94e0a3c236cb/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:24:42 crc kubenswrapper[4806]: I1204 05:24:42.789908 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7677974f-lk2tx_5e42fed2-6410-480b-ad51-7c8a5c383eea/dnsmasq-dns/0.log" Dec 04 05:24:42 crc kubenswrapper[4806]: I1204 05:24:42.962617 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0f1138ab-7d74-4dc5-883c-e32dd1f21546/glance-httpd/0.log" Dec 04 05:24:43 crc kubenswrapper[4806]: I1204 05:24:43.042078 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0f1138ab-7d74-4dc5-883c-e32dd1f21546/glance-log/0.log" Dec 04 05:24:43 crc kubenswrapper[4806]: I1204 05:24:43.322048 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2zzsv" event={"ID":"3c5e571a-32a2-401e-9370-af762975718f","Type":"ContainerStarted","Data":"facbc15edc54c25e03eaec86739c499362d6d0fac796989251d2bb409696eddb"} Dec 04 05:24:43 crc kubenswrapper[4806]: I1204 05:24:43.324006 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf8px" event={"ID":"c044b665-2475-438c-9616-f8856947509a","Type":"ContainerStarted","Data":"e2be315699dd3ef3f8a8d3b8831d1af1f509a1e674c6b969d311ac534a419d1b"} Dec 04 05:24:43 crc kubenswrapper[4806]: I1204 05:24:43.344661 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2zzsv" podStartSLOduration=2.886968434 podStartE2EDuration="7.344644938s" podCreationTimestamp="2025-12-04 05:24:36 +0000 UTC" firstStartedPulling="2025-12-04 05:24:38.250655514 +0000 UTC m=+5393.109168432" lastFinishedPulling="2025-12-04 05:24:42.708332008 +0000 UTC m=+5397.566844936" observedRunningTime="2025-12-04 05:24:43.342123259 +0000 UTC m=+5398.200636187" watchObservedRunningTime="2025-12-04 05:24:43.344644938 +0000 UTC m=+5398.203157866" Dec 04 05:24:43 crc kubenswrapper[4806]: I1204 05:24:43.368962 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hf8px" podStartSLOduration=2.8476347889999998 podStartE2EDuration="7.368945947s" podCreationTimestamp="2025-12-04 05:24:36 +0000 UTC" firstStartedPulling="2025-12-04 05:24:38.254514306 +0000 UTC m=+5393.113027234" lastFinishedPulling="2025-12-04 05:24:42.775825464 +0000 UTC m=+5397.634338392" observedRunningTime="2025-12-04 05:24:43.36742673 +0000 UTC m=+5398.225939658" watchObservedRunningTime="2025-12-04 05:24:43.368945947 +0000 UTC m=+5398.227458875" Dec 04 05:24:43 crc kubenswrapper[4806]: I1204 05:24:43.374574 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_5d515cfb-b26c-430f-9ee9-68b7d1a87fb6/glance-log/0.log" Dec 04 05:24:43 crc kubenswrapper[4806]: I1204 05:24:43.417306 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_5d515cfb-b26c-430f-9ee9-68b7d1a87fb6/glance-httpd/0.log" Dec 04 05:24:43 crc kubenswrapper[4806]: I1204 05:24:43.665751 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d8bd6c96d-4sqhf_d2d53864-c3c8-4e40-8791-0bac49ba561e/horizon/2.log" Dec 04 05:24:43 crc kubenswrapper[4806]: I1204 05:24:43.767178 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d8bd6c96d-4sqhf_d2d53864-c3c8-4e40-8791-0bac49ba561e/horizon/1.log" Dec 04 05:24:44 crc kubenswrapper[4806]: I1204 05:24:44.163496 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-n5w6j_f3284126-2542-43fc-8e96-829f03db14bc/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:24:44 crc kubenswrapper[4806]: I1204 05:24:44.234829 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-fdr25_3a53f8ff-b197-4f2b-8194-6424a21dae22/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:24:44 crc kubenswrapper[4806]: I1204 05:24:44.244269 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d8bd6c96d-4sqhf_d2d53864-c3c8-4e40-8791-0bac49ba561e/horizon-log/0.log" Dec 04 05:24:44 crc kubenswrapper[4806]: I1204 05:24:44.573305 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29413741-vfktv_106a7dfc-57ef-4177-821d-f77a0329ac90/keystone-cron/0.log" Dec 04 05:24:45 crc kubenswrapper[4806]: I1204 05:24:45.055521 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-64cc7bf6f4-fcxhm_0ffaf1c4-60c6-4d4d-b776-d2d23d2a99c5/keystone-api/0.log" Dec 04 05:24:45 crc kubenswrapper[4806]: I1204 05:24:45.114126 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_00dc1c6d-f6cc-49ac-935f-24007536925c/kube-state-metrics/0.log" Dec 04 05:24:45 crc kubenswrapper[4806]: I1204 05:24:45.236431 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-tzvr2_daec7717-fd5a-47be-a766-1ef8bbabab14/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:24:45 crc kubenswrapper[4806]: I1204 05:24:45.673274 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-258qq_4b974095-c38a-43c6-a6ac-67497d4f509f/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:24:45 crc kubenswrapper[4806]: I1204 05:24:45.982769 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-669cc9c8df-p88wk_7fbb687c-667a-4025-bd3c-1eb037758c0d/neutron-httpd/0.log" Dec 04 05:24:46 crc kubenswrapper[4806]: I1204 05:24:46.134058 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-669cc9c8df-p88wk_7fbb687c-667a-4025-bd3c-1eb037758c0d/neutron-api/0.log" Dec 04 05:24:46 crc kubenswrapper[4806]: I1204 05:24:46.735074 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:46 crc kubenswrapper[4806]: I1204 05:24:46.735382 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:46 crc kubenswrapper[4806]: I1204 05:24:46.808793 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_713e59da-b517-4099-bf3e-5be89d93e82c/nova-cell0-conductor-conductor/0.log" Dec 04 05:24:46 crc kubenswrapper[4806]: I1204 05:24:46.960865 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:24:46 crc kubenswrapper[4806]: I1204 05:24:46.960899 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:24:47 crc kubenswrapper[4806]: I1204 05:24:47.026479 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_e7137b5c-7316-46a0-abda-976ee89ac4ac/nova-cell1-conductor-conductor/0.log" Dec 04 05:24:47 crc kubenswrapper[4806]: I1204 05:24:47.463088 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4d149e6e-bdab-4137-a48a-27ed21680ab8/nova-api-log/0.log" Dec 04 05:24:47 crc kubenswrapper[4806]: I1204 05:24:47.528406 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_3999d3e5-7b66-4792-9fd0-996d0d97efd2/nova-cell1-novncproxy-novncproxy/0.log" Dec 04 05:24:47 crc kubenswrapper[4806]: I1204 05:24:47.660602 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-l4hb8_21f3ccbb-669c-4055-b85c-f0c69fe61a8e/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:24:47 crc kubenswrapper[4806]: I1204 05:24:47.800289 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-2zzsv" podUID="3c5e571a-32a2-401e-9370-af762975718f" containerName="registry-server" probeResult="failure" output=< Dec 04 05:24:47 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Dec 04 05:24:47 crc kubenswrapper[4806]: > Dec 04 05:24:47 crc kubenswrapper[4806]: I1204 05:24:47.935265 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac/nova-metadata-log/0.log" Dec 04 05:24:48 crc kubenswrapper[4806]: I1204 05:24:48.014434 4806 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-hf8px" podUID="c044b665-2475-438c-9616-f8856947509a" containerName="registry-server" probeResult="failure" output=< Dec 04 05:24:48 crc kubenswrapper[4806]: timeout: failed to connect service ":50051" within 1s Dec 04 05:24:48 crc kubenswrapper[4806]: > Dec 04 05:24:48 crc kubenswrapper[4806]: I1204 05:24:48.023129 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4d149e6e-bdab-4137-a48a-27ed21680ab8/nova-api-api/0.log" Dec 04 05:24:48 crc kubenswrapper[4806]: I1204 05:24:48.504853 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0d1a2b83-04d7-4db8-acd1-456cc0c33206/mysql-bootstrap/0.log" Dec 04 05:24:48 crc kubenswrapper[4806]: I1204 05:24:48.612781 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0d1a2b83-04d7-4db8-acd1-456cc0c33206/mysql-bootstrap/0.log" Dec 04 05:24:48 crc kubenswrapper[4806]: I1204 05:24:48.766641 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_0d1a2b83-04d7-4db8-acd1-456cc0c33206/galera/0.log" Dec 04 05:24:48 crc kubenswrapper[4806]: I1204 05:24:48.829952 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_a32a7b4e-8881-437b-843e-54a6d1e5c11e/nova-scheduler-scheduler/0.log" Dec 04 05:24:48 crc kubenswrapper[4806]: I1204 05:24:48.982329 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_99326fd5-69d5-4e7d-b5c6-f7808344ad6d/mysql-bootstrap/0.log" Dec 04 05:24:49 crc kubenswrapper[4806]: I1204 05:24:49.248315 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_99326fd5-69d5-4e7d-b5c6-f7808344ad6d/mysql-bootstrap/0.log" Dec 04 05:24:49 crc kubenswrapper[4806]: I1204 05:24:49.278058 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_99326fd5-69d5-4e7d-b5c6-f7808344ad6d/galera/0.log" Dec 04 05:24:49 crc kubenswrapper[4806]: I1204 05:24:49.427390 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_266dbc23-c65d-4b9d-978d-09f9de4a06e5/openstackclient/0.log" Dec 04 05:24:49 crc kubenswrapper[4806]: I1204 05:24:49.595153 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-4kl52_0b91c83f-0036-42a8-97b4-86846f14466a/openstack-network-exporter/0.log" Dec 04 05:24:49 crc kubenswrapper[4806]: I1204 05:24:49.826993 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2sz9t_63e4ca21-d6b3-4e93-965e-3dde5e35027f/ovsdb-server-init/0.log" Dec 04 05:24:50 crc kubenswrapper[4806]: I1204 05:24:50.035416 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2sz9t_63e4ca21-d6b3-4e93-965e-3dde5e35027f/ovsdb-server-init/0.log" Dec 04 05:24:50 crc kubenswrapper[4806]: I1204 05:24:50.049824 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2sz9t_63e4ca21-d6b3-4e93-965e-3dde5e35027f/ovsdb-server/0.log" Dec 04 05:24:50 crc kubenswrapper[4806]: I1204 05:24:50.130679 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-2sz9t_63e4ca21-d6b3-4e93-965e-3dde5e35027f/ovs-vswitchd/0.log" Dec 04 05:24:50 crc kubenswrapper[4806]: I1204 05:24:50.272994 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c8e58d37-31b1-476b-9eb1-e2f0a2b2eeac/nova-metadata-metadata/0.log" Dec 04 05:24:50 crc kubenswrapper[4806]: I1204 05:24:50.284656 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-tx487_ef210587-097b-48d8-8097-dc1b95e2456e/ovn-controller/0.log" Dec 04 05:24:50 crc kubenswrapper[4806]: I1204 05:24:50.526684 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_55f867c0-7de7-40bc-839a-4f9d9d09dd2e/openstack-network-exporter/0.log" Dec 04 05:24:50 crc kubenswrapper[4806]: I1204 05:24:50.642987 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-mbhj2_b4d109a9-916a-41b6-9008-15291d44a0a4/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:24:50 crc kubenswrapper[4806]: I1204 05:24:50.717241 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_55f867c0-7de7-40bc-839a-4f9d9d09dd2e/ovn-northd/0.log" Dec 04 05:24:50 crc kubenswrapper[4806]: I1204 05:24:50.856538 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_75e0b5e3-e994-4ef3-9673-539ca96485c5/openstack-network-exporter/0.log" Dec 04 05:24:50 crc kubenswrapper[4806]: I1204 05:24:50.895493 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_75e0b5e3-e994-4ef3-9673-539ca96485c5/ovsdbserver-nb/0.log" Dec 04 05:24:51 crc kubenswrapper[4806]: I1204 05:24:51.072619 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_09739947-6f7d-4a4e-8a85-05a8a7345894/openstack-network-exporter/0.log" Dec 04 05:24:51 crc kubenswrapper[4806]: I1204 05:24:51.111984 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_09739947-6f7d-4a4e-8a85-05a8a7345894/ovsdbserver-sb/0.log" Dec 04 05:24:51 crc kubenswrapper[4806]: I1204 05:24:51.546044 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_17d9cfc5-f7f9-442f-8174-2b61041b8f70/setup-container/0.log" Dec 04 05:24:51 crc kubenswrapper[4806]: I1204 05:24:51.558607 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-d9c876cd8-zfctt_122ae99a-2891-4bdb-903a-9cccfb3df2a1/placement-api/0.log" Dec 04 05:24:51 crc kubenswrapper[4806]: I1204 05:24:51.607470 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-d9c876cd8-zfctt_122ae99a-2891-4bdb-903a-9cccfb3df2a1/placement-log/0.log" Dec 04 05:24:51 crc kubenswrapper[4806]: I1204 05:24:51.726642 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_17d9cfc5-f7f9-442f-8174-2b61041b8f70/setup-container/0.log" Dec 04 05:24:51 crc kubenswrapper[4806]: I1204 05:24:51.836427 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_17d9cfc5-f7f9-442f-8174-2b61041b8f70/rabbitmq/0.log" Dec 04 05:24:51 crc kubenswrapper[4806]: I1204 05:24:51.929581 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5a2a4a0f-f385-4fad-9825-2824f47dd8db/setup-container/0.log" Dec 04 05:24:52 crc kubenswrapper[4806]: I1204 05:24:52.134412 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5a2a4a0f-f385-4fad-9825-2824f47dd8db/setup-container/0.log" Dec 04 05:24:52 crc kubenswrapper[4806]: I1204 05:24:52.157667 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5a2a4a0f-f385-4fad-9825-2824f47dd8db/rabbitmq/0.log" Dec 04 05:24:52 crc kubenswrapper[4806]: I1204 05:24:52.216471 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-fprln_a74530eb-f7f4-40f7-b1fc-c66cba904ce5/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:24:52 crc kubenswrapper[4806]: I1204 05:24:52.421659 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-ds7xn_ef28af41-0fb7-4e9a-adbe-c08fdf320e95/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:24:52 crc kubenswrapper[4806]: I1204 05:24:52.557490 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-b5wld_410d8040-e0b5-4869-bb16-41e5b59487d3/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:24:52 crc kubenswrapper[4806]: I1204 05:24:52.750973 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-25667_e5f894f3-1fed-46cc-9caa-8f61712e8ab3/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:24:52 crc kubenswrapper[4806]: I1204 05:24:52.855906 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-cddcl_cf2d196e-c288-449d-a106-2ca0b6697582/ssh-known-hosts-edpm-deployment/0.log" Dec 04 05:24:53 crc kubenswrapper[4806]: I1204 05:24:53.048838 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-685c486f6f-6zmmj_794d2127-4482-4196-b558-d24f003e96ca/proxy-server/0.log" Dec 04 05:24:53 crc kubenswrapper[4806]: I1204 05:24:53.221085 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-t7k9p_85f9dc59-e013-464d-b305-77c85cb82f4a/swift-ring-rebalance/0.log" Dec 04 05:24:53 crc kubenswrapper[4806]: I1204 05:24:53.275610 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-685c486f6f-6zmmj_794d2127-4482-4196-b558-d24f003e96ca/proxy-httpd/0.log" Dec 04 05:24:53 crc kubenswrapper[4806]: I1204 05:24:53.407905 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/account-auditor/0.log" Dec 04 05:24:53 crc kubenswrapper[4806]: I1204 05:24:53.456466 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/account-reaper/0.log" Dec 04 05:24:53 crc kubenswrapper[4806]: I1204 05:24:53.595616 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/account-server/0.log" Dec 04 05:24:53 crc kubenswrapper[4806]: I1204 05:24:53.703714 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/account-replicator/0.log" Dec 04 05:24:53 crc kubenswrapper[4806]: I1204 05:24:53.737098 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/container-auditor/0.log" Dec 04 05:24:53 crc kubenswrapper[4806]: I1204 05:24:53.740877 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/container-replicator/0.log" Dec 04 05:24:53 crc kubenswrapper[4806]: I1204 05:24:53.853733 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/container-server/0.log" Dec 04 05:24:53 crc kubenswrapper[4806]: I1204 05:24:53.917825 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/container-updater/0.log" Dec 04 05:24:54 crc kubenswrapper[4806]: I1204 05:24:54.005882 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/object-expirer/0.log" Dec 04 05:24:54 crc kubenswrapper[4806]: I1204 05:24:54.011127 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/object-auditor/0.log" Dec 04 05:24:54 crc kubenswrapper[4806]: I1204 05:24:54.153779 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/object-server/0.log" Dec 04 05:24:54 crc kubenswrapper[4806]: I1204 05:24:54.197762 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/object-replicator/0.log" Dec 04 05:24:54 crc kubenswrapper[4806]: I1204 05:24:54.317649 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/object-updater/0.log" Dec 04 05:24:54 crc kubenswrapper[4806]: I1204 05:24:54.322294 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/rsync/0.log" Dec 04 05:24:54 crc kubenswrapper[4806]: I1204 05:24:54.441193 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_44220187-6c1e-4755-8420-d629dd59a84a/swift-recon-cron/0.log" Dec 04 05:24:54 crc kubenswrapper[4806]: I1204 05:24:54.735829 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-stpm5_c9247c4e-1c6a-4110-8d97-6829c6064d48/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:24:54 crc kubenswrapper[4806]: I1204 05:24:54.743371 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_319a7357-70c6-4b54-8e97-9860dd0618e7/tempest-tests-tempest-tests-runner/0.log" Dec 04 05:24:54 crc kubenswrapper[4806]: I1204 05:24:54.939990 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_4922f2b6-05e3-4f78-a14a-46ca2668edfb/test-operator-logs-container/0.log" Dec 04 05:24:54 crc kubenswrapper[4806]: I1204 05:24:54.972984 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-phq5h_02021c61-fa0c-4aea-80d4-b0eb592c0d06/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 04 05:24:56 crc kubenswrapper[4806]: I1204 05:24:56.834812 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:56 crc kubenswrapper[4806]: I1204 05:24:56.883102 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:57 crc kubenswrapper[4806]: I1204 05:24:57.015605 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:24:57 crc kubenswrapper[4806]: I1204 05:24:57.048488 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:24:57 crc kubenswrapper[4806]: I1204 05:24:57.048545 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:24:57 crc kubenswrapper[4806]: I1204 05:24:57.071097 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:24:57 crc kubenswrapper[4806]: I1204 05:24:57.072829 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2zzsv"] Dec 04 05:24:58 crc kubenswrapper[4806]: I1204 05:24:58.451784 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2zzsv" podUID="3c5e571a-32a2-401e-9370-af762975718f" containerName="registry-server" containerID="cri-o://facbc15edc54c25e03eaec86739c499362d6d0fac796989251d2bb409696eddb" gracePeriod=2 Dec 04 05:24:58 crc kubenswrapper[4806]: I1204 05:24:58.976404 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.071899 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5e571a-32a2-401e-9370-af762975718f-utilities\") pod \"3c5e571a-32a2-401e-9370-af762975718f\" (UID: \"3c5e571a-32a2-401e-9370-af762975718f\") " Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.072089 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5e571a-32a2-401e-9370-af762975718f-catalog-content\") pod \"3c5e571a-32a2-401e-9370-af762975718f\" (UID: \"3c5e571a-32a2-401e-9370-af762975718f\") " Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.072148 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhzxf\" (UniqueName: \"kubernetes.io/projected/3c5e571a-32a2-401e-9370-af762975718f-kube-api-access-nhzxf\") pod \"3c5e571a-32a2-401e-9370-af762975718f\" (UID: \"3c5e571a-32a2-401e-9370-af762975718f\") " Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.072890 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c5e571a-32a2-401e-9370-af762975718f-utilities" (OuterVolumeSpecName: "utilities") pod "3c5e571a-32a2-401e-9370-af762975718f" (UID: "3c5e571a-32a2-401e-9370-af762975718f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.094365 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c5e571a-32a2-401e-9370-af762975718f-kube-api-access-nhzxf" (OuterVolumeSpecName: "kube-api-access-nhzxf") pod "3c5e571a-32a2-401e-9370-af762975718f" (UID: "3c5e571a-32a2-401e-9370-af762975718f"). InnerVolumeSpecName "kube-api-access-nhzxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.144785 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c5e571a-32a2-401e-9370-af762975718f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c5e571a-32a2-401e-9370-af762975718f" (UID: "3c5e571a-32a2-401e-9370-af762975718f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.173726 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhzxf\" (UniqueName: \"kubernetes.io/projected/3c5e571a-32a2-401e-9370-af762975718f-kube-api-access-nhzxf\") on node \"crc\" DevicePath \"\"" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.173761 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c5e571a-32a2-401e-9370-af762975718f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.173776 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c5e571a-32a2-401e-9370-af762975718f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.283304 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hf8px"] Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.286544 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hf8px" podUID="c044b665-2475-438c-9616-f8856947509a" containerName="registry-server" containerID="cri-o://e2be315699dd3ef3f8a8d3b8831d1af1f509a1e674c6b969d311ac534a419d1b" gracePeriod=2 Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.467646 4806 generic.go:334] "Generic (PLEG): container finished" podID="3c5e571a-32a2-401e-9370-af762975718f" containerID="facbc15edc54c25e03eaec86739c499362d6d0fac796989251d2bb409696eddb" exitCode=0 Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.467745 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2zzsv" event={"ID":"3c5e571a-32a2-401e-9370-af762975718f","Type":"ContainerDied","Data":"facbc15edc54c25e03eaec86739c499362d6d0fac796989251d2bb409696eddb"} Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.467774 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2zzsv" event={"ID":"3c5e571a-32a2-401e-9370-af762975718f","Type":"ContainerDied","Data":"291b71357975a2e0e8c3231728b1853dc5888f16fabfd14adefa98a6299d57cb"} Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.467806 4806 scope.go:117] "RemoveContainer" containerID="facbc15edc54c25e03eaec86739c499362d6d0fac796989251d2bb409696eddb" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.467956 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2zzsv" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.491191 4806 generic.go:334] "Generic (PLEG): container finished" podID="c044b665-2475-438c-9616-f8856947509a" containerID="e2be315699dd3ef3f8a8d3b8831d1af1f509a1e674c6b969d311ac534a419d1b" exitCode=0 Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.491253 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf8px" event={"ID":"c044b665-2475-438c-9616-f8856947509a","Type":"ContainerDied","Data":"e2be315699dd3ef3f8a8d3b8831d1af1f509a1e674c6b969d311ac534a419d1b"} Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.511424 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2zzsv"] Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.528054 4806 scope.go:117] "RemoveContainer" containerID="f6441f48494b257e10f5db8abc4b6d31e7a028de70781b9b54b7331146dd87ce" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.540389 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2zzsv"] Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.552823 4806 scope.go:117] "RemoveContainer" containerID="5eb53285a1df641afad87d6d04ea0419da44b734d5f8be63547c6dae6a91faef" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.594083 4806 scope.go:117] "RemoveContainer" containerID="facbc15edc54c25e03eaec86739c499362d6d0fac796989251d2bb409696eddb" Dec 04 05:24:59 crc kubenswrapper[4806]: E1204 05:24:59.594489 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"facbc15edc54c25e03eaec86739c499362d6d0fac796989251d2bb409696eddb\": container with ID starting with facbc15edc54c25e03eaec86739c499362d6d0fac796989251d2bb409696eddb not found: ID does not exist" containerID="facbc15edc54c25e03eaec86739c499362d6d0fac796989251d2bb409696eddb" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.594519 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"facbc15edc54c25e03eaec86739c499362d6d0fac796989251d2bb409696eddb"} err="failed to get container status \"facbc15edc54c25e03eaec86739c499362d6d0fac796989251d2bb409696eddb\": rpc error: code = NotFound desc = could not find container \"facbc15edc54c25e03eaec86739c499362d6d0fac796989251d2bb409696eddb\": container with ID starting with facbc15edc54c25e03eaec86739c499362d6d0fac796989251d2bb409696eddb not found: ID does not exist" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.594538 4806 scope.go:117] "RemoveContainer" containerID="f6441f48494b257e10f5db8abc4b6d31e7a028de70781b9b54b7331146dd87ce" Dec 04 05:24:59 crc kubenswrapper[4806]: E1204 05:24:59.598068 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6441f48494b257e10f5db8abc4b6d31e7a028de70781b9b54b7331146dd87ce\": container with ID starting with f6441f48494b257e10f5db8abc4b6d31e7a028de70781b9b54b7331146dd87ce not found: ID does not exist" containerID="f6441f48494b257e10f5db8abc4b6d31e7a028de70781b9b54b7331146dd87ce" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.598126 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6441f48494b257e10f5db8abc4b6d31e7a028de70781b9b54b7331146dd87ce"} err="failed to get container status \"f6441f48494b257e10f5db8abc4b6d31e7a028de70781b9b54b7331146dd87ce\": rpc error: code = NotFound desc = could not find container \"f6441f48494b257e10f5db8abc4b6d31e7a028de70781b9b54b7331146dd87ce\": container with ID starting with f6441f48494b257e10f5db8abc4b6d31e7a028de70781b9b54b7331146dd87ce not found: ID does not exist" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.598160 4806 scope.go:117] "RemoveContainer" containerID="5eb53285a1df641afad87d6d04ea0419da44b734d5f8be63547c6dae6a91faef" Dec 04 05:24:59 crc kubenswrapper[4806]: E1204 05:24:59.618167 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5eb53285a1df641afad87d6d04ea0419da44b734d5f8be63547c6dae6a91faef\": container with ID starting with 5eb53285a1df641afad87d6d04ea0419da44b734d5f8be63547c6dae6a91faef not found: ID does not exist" containerID="5eb53285a1df641afad87d6d04ea0419da44b734d5f8be63547c6dae6a91faef" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.619431 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5eb53285a1df641afad87d6d04ea0419da44b734d5f8be63547c6dae6a91faef"} err="failed to get container status \"5eb53285a1df641afad87d6d04ea0419da44b734d5f8be63547c6dae6a91faef\": rpc error: code = NotFound desc = could not find container \"5eb53285a1df641afad87d6d04ea0419da44b734d5f8be63547c6dae6a91faef\": container with ID starting with 5eb53285a1df641afad87d6d04ea0419da44b734d5f8be63547c6dae6a91faef not found: ID does not exist" Dec 04 05:24:59 crc kubenswrapper[4806]: I1204 05:24:59.887813 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:25:00 crc kubenswrapper[4806]: I1204 05:25:00.012642 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmrxk\" (UniqueName: \"kubernetes.io/projected/c044b665-2475-438c-9616-f8856947509a-kube-api-access-zmrxk\") pod \"c044b665-2475-438c-9616-f8856947509a\" (UID: \"c044b665-2475-438c-9616-f8856947509a\") " Dec 04 05:25:00 crc kubenswrapper[4806]: I1204 05:25:00.012811 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c044b665-2475-438c-9616-f8856947509a-utilities\") pod \"c044b665-2475-438c-9616-f8856947509a\" (UID: \"c044b665-2475-438c-9616-f8856947509a\") " Dec 04 05:25:00 crc kubenswrapper[4806]: I1204 05:25:00.012864 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c044b665-2475-438c-9616-f8856947509a-catalog-content\") pod \"c044b665-2475-438c-9616-f8856947509a\" (UID: \"c044b665-2475-438c-9616-f8856947509a\") " Dec 04 05:25:00 crc kubenswrapper[4806]: I1204 05:25:00.013859 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c044b665-2475-438c-9616-f8856947509a-utilities" (OuterVolumeSpecName: "utilities") pod "c044b665-2475-438c-9616-f8856947509a" (UID: "c044b665-2475-438c-9616-f8856947509a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:25:00 crc kubenswrapper[4806]: I1204 05:25:00.019232 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c044b665-2475-438c-9616-f8856947509a-kube-api-access-zmrxk" (OuterVolumeSpecName: "kube-api-access-zmrxk") pod "c044b665-2475-438c-9616-f8856947509a" (UID: "c044b665-2475-438c-9616-f8856947509a"). InnerVolumeSpecName "kube-api-access-zmrxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:25:00 crc kubenswrapper[4806]: I1204 05:25:00.114469 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c044b665-2475-438c-9616-f8856947509a-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 05:25:00 crc kubenswrapper[4806]: I1204 05:25:00.114501 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmrxk\" (UniqueName: \"kubernetes.io/projected/c044b665-2475-438c-9616-f8856947509a-kube-api-access-zmrxk\") on node \"crc\" DevicePath \"\"" Dec 04 05:25:00 crc kubenswrapper[4806]: I1204 05:25:00.115390 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c044b665-2475-438c-9616-f8856947509a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c044b665-2475-438c-9616-f8856947509a" (UID: "c044b665-2475-438c-9616-f8856947509a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:25:00 crc kubenswrapper[4806]: I1204 05:25:00.215626 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c044b665-2475-438c-9616-f8856947509a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 05:25:00 crc kubenswrapper[4806]: I1204 05:25:00.503480 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hf8px" event={"ID":"c044b665-2475-438c-9616-f8856947509a","Type":"ContainerDied","Data":"c119a31ae2c73abb1509e62666e49f31bf7b91d621dae22c95abe5b77d8ce246"} Dec 04 05:25:00 crc kubenswrapper[4806]: I1204 05:25:00.503536 4806 scope.go:117] "RemoveContainer" containerID="e2be315699dd3ef3f8a8d3b8831d1af1f509a1e674c6b969d311ac534a419d1b" Dec 04 05:25:00 crc kubenswrapper[4806]: I1204 05:25:00.503972 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hf8px" Dec 04 05:25:00 crc kubenswrapper[4806]: I1204 05:25:00.524632 4806 scope.go:117] "RemoveContainer" containerID="b82e25d3ddcad7fb5643bfe3aebde7e140f5b309bde0b3a5a0a06232538d317f" Dec 04 05:25:00 crc kubenswrapper[4806]: I1204 05:25:00.544198 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hf8px"] Dec 04 05:25:00 crc kubenswrapper[4806]: I1204 05:25:00.555999 4806 scope.go:117] "RemoveContainer" containerID="bdf9a16d6d82d2142ab9ac9771f5a70211a71c8d71df85a13e985b680438d26c" Dec 04 05:25:00 crc kubenswrapper[4806]: I1204 05:25:00.579946 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hf8px"] Dec 04 05:25:01 crc kubenswrapper[4806]: I1204 05:25:01.432759 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c5e571a-32a2-401e-9370-af762975718f" path="/var/lib/kubelet/pods/3c5e571a-32a2-401e-9370-af762975718f/volumes" Dec 04 05:25:01 crc kubenswrapper[4806]: I1204 05:25:01.433731 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c044b665-2475-438c-9616-f8856947509a" path="/var/lib/kubelet/pods/c044b665-2475-438c-9616-f8856947509a/volumes" Dec 04 05:25:08 crc kubenswrapper[4806]: I1204 05:25:08.314078 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_0face013-e59b-4e32-ace5-d64d45302e7c/memcached/0.log" Dec 04 05:25:26 crc kubenswrapper[4806]: I1204 05:25:26.251915 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh_a72e0bcd-e604-4243-b294-70ebe7ba990c/util/0.log" Dec 04 05:25:26 crc kubenswrapper[4806]: I1204 05:25:26.514916 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh_a72e0bcd-e604-4243-b294-70ebe7ba990c/util/0.log" Dec 04 05:25:26 crc kubenswrapper[4806]: I1204 05:25:26.572310 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh_a72e0bcd-e604-4243-b294-70ebe7ba990c/pull/0.log" Dec 04 05:25:26 crc kubenswrapper[4806]: I1204 05:25:26.594044 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh_a72e0bcd-e604-4243-b294-70ebe7ba990c/pull/0.log" Dec 04 05:25:26 crc kubenswrapper[4806]: I1204 05:25:26.749973 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh_a72e0bcd-e604-4243-b294-70ebe7ba990c/util/0.log" Dec 04 05:25:26 crc kubenswrapper[4806]: I1204 05:25:26.750700 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh_a72e0bcd-e604-4243-b294-70ebe7ba990c/pull/0.log" Dec 04 05:25:26 crc kubenswrapper[4806]: I1204 05:25:26.811779 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02cjhlh_a72e0bcd-e604-4243-b294-70ebe7ba990c/extract/0.log" Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.005296 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-g2dfw_b2bba37e-d570-432f-9520-9b5de2005e5c/kube-rbac-proxy/0.log" Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.046666 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.046714 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.046759 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.047469 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a025afa4e0f2bafc160c671d65ca45902522db3c999cd6d04dd15c40321dcf4a"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.047518 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://a025afa4e0f2bafc160c671d65ca45902522db3c999cd6d04dd15c40321dcf4a" gracePeriod=600 Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.084701 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-g2dfw_b2bba37e-d570-432f-9520-9b5de2005e5c/manager/0.log" Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.243048 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-25nxx_a022d699-4a2b-49e2-8d8f-787a3d888df8/kube-rbac-proxy/0.log" Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.334535 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-25nxx_a022d699-4a2b-49e2-8d8f-787a3d888df8/manager/0.log" Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.460583 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-mxmf2_91ab6e01-027e-4e7a-a250-374b4e8873fc/kube-rbac-proxy/0.log" Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.517816 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-mxmf2_91ab6e01-027e-4e7a-a250-374b4e8873fc/manager/0.log" Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.697294 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-xd6pj_7213e0bc-21af-4df7-929d-f915e343f184/kube-rbac-proxy/0.log" Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.769170 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="a025afa4e0f2bafc160c671d65ca45902522db3c999cd6d04dd15c40321dcf4a" exitCode=0 Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.769232 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"a025afa4e0f2bafc160c671d65ca45902522db3c999cd6d04dd15c40321dcf4a"} Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.769308 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerStarted","Data":"cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19"} Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.769325 4806 scope.go:117] "RemoveContainer" containerID="b0c1c9e002aa995f163f32495f63863a41a2507a6c7ea4cea86ee4dd47c98459" Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.782513 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-xd6pj_7213e0bc-21af-4df7-929d-f915e343f184/manager/0.log" Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.970764 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-vhxjt_b496f26d-2493-416f-a0f8-5b33e41dbda0/kube-rbac-proxy/0.log" Dec 04 05:25:27 crc kubenswrapper[4806]: I1204 05:25:27.990277 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-vhxjt_b496f26d-2493-416f-a0f8-5b33e41dbda0/manager/0.log" Dec 04 05:25:28 crc kubenswrapper[4806]: I1204 05:25:28.177815 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-vtcg5_084c526b-94d8-43ae-9ffb-33aa215de8c2/kube-rbac-proxy/0.log" Dec 04 05:25:28 crc kubenswrapper[4806]: I1204 05:25:28.184624 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-vtcg5_084c526b-94d8-43ae-9ffb-33aa215de8c2/manager/0.log" Dec 04 05:25:28 crc kubenswrapper[4806]: I1204 05:25:28.323113 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-g6svq_04d2c949-c17a-4299-a9be-b56a7c34866c/kube-rbac-proxy/0.log" Dec 04 05:25:28 crc kubenswrapper[4806]: I1204 05:25:28.367867 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-f6k5f_5f55ddcf-ca5d-4249-9384-e541ae7d4536/kube-rbac-proxy/0.log" Dec 04 05:25:28 crc kubenswrapper[4806]: I1204 05:25:28.537842 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-g6svq_04d2c949-c17a-4299-a9be-b56a7c34866c/manager/0.log" Dec 04 05:25:28 crc kubenswrapper[4806]: I1204 05:25:28.561074 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-f6k5f_5f55ddcf-ca5d-4249-9384-e541ae7d4536/manager/0.log" Dec 04 05:25:28 crc kubenswrapper[4806]: I1204 05:25:28.720090 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-nt5vq_669dd56a-69b0-4315-aacc-c8f085f0b76a/kube-rbac-proxy/0.log" Dec 04 05:25:28 crc kubenswrapper[4806]: I1204 05:25:28.859756 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-nt5vq_669dd56a-69b0-4315-aacc-c8f085f0b76a/manager/0.log" Dec 04 05:25:28 crc kubenswrapper[4806]: I1204 05:25:28.950017 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-2spbg_66428f64-e34e-4b37-b911-13e7142f1c18/kube-rbac-proxy/0.log" Dec 04 05:25:29 crc kubenswrapper[4806]: I1204 05:25:29.024388 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-2spbg_66428f64-e34e-4b37-b911-13e7142f1c18/manager/0.log" Dec 04 05:25:29 crc kubenswrapper[4806]: I1204 05:25:29.115546 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-45jzr_310d58eb-6345-4cba-8a01-373c65c7bc40/kube-rbac-proxy/0.log" Dec 04 05:25:29 crc kubenswrapper[4806]: I1204 05:25:29.187627 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-45jzr_310d58eb-6345-4cba-8a01-373c65c7bc40/manager/0.log" Dec 04 05:25:29 crc kubenswrapper[4806]: I1204 05:25:29.427308 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-2b2qk_95ae1fa1-3ac7-4040-86da-150daca97f6b/kube-rbac-proxy/0.log" Dec 04 05:25:29 crc kubenswrapper[4806]: I1204 05:25:29.500109 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-2b2qk_95ae1fa1-3ac7-4040-86da-150daca97f6b/manager/0.log" Dec 04 05:25:29 crc kubenswrapper[4806]: I1204 05:25:29.627632 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-pk7ws_5f479a93-bb3a-45b8-aa74-60e9f0a7aebd/kube-rbac-proxy/0.log" Dec 04 05:25:29 crc kubenswrapper[4806]: I1204 05:25:29.732042 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-pk7ws_5f479a93-bb3a-45b8-aa74-60e9f0a7aebd/manager/0.log" Dec 04 05:25:29 crc kubenswrapper[4806]: I1204 05:25:29.772528 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-kcmvz_1cfe7f94-e830-4145-a3e7-8a3f694adda1/kube-rbac-proxy/0.log" Dec 04 05:25:29 crc kubenswrapper[4806]: I1204 05:25:29.873754 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-kcmvz_1cfe7f94-e830-4145-a3e7-8a3f694adda1/manager/0.log" Dec 04 05:25:29 crc kubenswrapper[4806]: I1204 05:25:29.979682 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7_2a7907e6-c36f-4534-9460-80c0091af286/kube-rbac-proxy/0.log" Dec 04 05:25:29 crc kubenswrapper[4806]: I1204 05:25:29.987530 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4q6js7_2a7907e6-c36f-4534-9460-80c0091af286/manager/0.log" Dec 04 05:25:30 crc kubenswrapper[4806]: I1204 05:25:30.346188 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7bfcb77fcc-bx765_4d229897-71a8-4b61-a9be-4d963e1384ac/operator/0.log" Dec 04 05:25:30 crc kubenswrapper[4806]: I1204 05:25:30.431001 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-gsvwx_89d5c50c-5d00-4f21-9e60-8c6692a0ad33/registry-server/0.log" Dec 04 05:25:30 crc kubenswrapper[4806]: I1204 05:25:30.622786 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-sktq6_e06aa87c-9f7d-44f7-b96d-27141d2fff93/kube-rbac-proxy/0.log" Dec 04 05:25:30 crc kubenswrapper[4806]: I1204 05:25:30.689005 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-sktq6_e06aa87c-9f7d-44f7-b96d-27141d2fff93/manager/0.log" Dec 04 05:25:30 crc kubenswrapper[4806]: I1204 05:25:30.835790 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-mnhlt_45192fbb-caf2-4df8-82a6-c222f73b033c/kube-rbac-proxy/0.log" Dec 04 05:25:31 crc kubenswrapper[4806]: I1204 05:25:31.023329 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-mnhlt_45192fbb-caf2-4df8-82a6-c222f73b033c/manager/0.log" Dec 04 05:25:31 crc kubenswrapper[4806]: I1204 05:25:31.131602 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-cwnth_be179b75-6e57-4537-b3ee-a57dc049948e/operator/0.log" Dec 04 05:25:31 crc kubenswrapper[4806]: I1204 05:25:31.271359 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8d698776b-npcz2_ff1a6781-908b-419f-8a02-10abbbfa0816/manager/0.log" Dec 04 05:25:31 crc kubenswrapper[4806]: I1204 05:25:31.275618 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-fdqbx_3580e441-1df2-4fe9-8969-e1621fc554dc/kube-rbac-proxy/0.log" Dec 04 05:25:31 crc kubenswrapper[4806]: I1204 05:25:31.344679 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-fdqbx_3580e441-1df2-4fe9-8969-e1621fc554dc/manager/0.log" Dec 04 05:25:31 crc kubenswrapper[4806]: I1204 05:25:31.428537 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-kvnjv_647621de-6ecf-4ebe-96f2-e85f26e4917e/kube-rbac-proxy/0.log" Dec 04 05:25:31 crc kubenswrapper[4806]: I1204 05:25:31.526142 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-kvnjv_647621de-6ecf-4ebe-96f2-e85f26e4917e/manager/0.log" Dec 04 05:25:31 crc kubenswrapper[4806]: I1204 05:25:31.575146 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-69l55_2473e4f1-90f0-43a0-9e72-0c761d335d76/kube-rbac-proxy/0.log" Dec 04 05:25:31 crc kubenswrapper[4806]: I1204 05:25:31.621721 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-69l55_2473e4f1-90f0-43a0-9e72-0c761d335d76/manager/0.log" Dec 04 05:25:31 crc kubenswrapper[4806]: I1204 05:25:31.746948 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-2lk62_a3827706-d53c-48f3-b4cf-1c240c0f868c/kube-rbac-proxy/0.log" Dec 04 05:25:31 crc kubenswrapper[4806]: I1204 05:25:31.778685 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-2lk62_a3827706-d53c-48f3-b4cf-1c240c0f868c/manager/0.log" Dec 04 05:25:50 crc kubenswrapper[4806]: I1204 05:25:50.865889 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-dgkrj_0647d456-31a3-4906-9d62-f43b4cfef5cb/control-plane-machine-set-operator/0.log" Dec 04 05:25:51 crc kubenswrapper[4806]: I1204 05:25:51.040693 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j8m72_7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6/kube-rbac-proxy/0.log" Dec 04 05:25:51 crc kubenswrapper[4806]: I1204 05:25:51.050995 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j8m72_7c7a0a70-8ee9-4f77-850e-e2a5c72ef8d6/machine-api-operator/0.log" Dec 04 05:26:04 crc kubenswrapper[4806]: I1204 05:26:04.095156 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-6l44d_ed32bab6-fee8-43d7-a157-9e8d5d17b98a/cert-manager-controller/0.log" Dec 04 05:26:04 crc kubenswrapper[4806]: I1204 05:26:04.294127 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-8vl9s_65faa3e8-8810-4a0e-a3ec-bb2b4d6283fb/cert-manager-cainjector/0.log" Dec 04 05:26:04 crc kubenswrapper[4806]: I1204 05:26:04.448458 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-dzbcz_7934a456-a3ec-4a88-8857-17d7f7db5837/cert-manager-webhook/0.log" Dec 04 05:26:17 crc kubenswrapper[4806]: I1204 05:26:17.257457 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-572xx_b0e181e1-d1fa-460f-963f-cbd235500b2e/nmstate-console-plugin/0.log" Dec 04 05:26:17 crc kubenswrapper[4806]: I1204 05:26:17.369055 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-pdg7x_843789a7-01bb-42c0-9f30-aa52af788147/nmstate-handler/0.log" Dec 04 05:26:17 crc kubenswrapper[4806]: I1204 05:26:17.440697 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-dlkzr_7b294b89-f409-40b7-8ba4-72e53da24a6f/nmstate-metrics/0.log" Dec 04 05:26:17 crc kubenswrapper[4806]: I1204 05:26:17.466346 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-dlkzr_7b294b89-f409-40b7-8ba4-72e53da24a6f/kube-rbac-proxy/0.log" Dec 04 05:26:17 crc kubenswrapper[4806]: I1204 05:26:17.659197 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-zz62j_93c65cc3-ea7f-4f77-af33-8acdcd03432c/nmstate-operator/0.log" Dec 04 05:26:17 crc kubenswrapper[4806]: I1204 05:26:17.691750 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-q9x68_eae97ef2-3aeb-4303-b26f-0f84a75abda0/nmstate-webhook/0.log" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.026910 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7tgg8"] Dec 04 05:26:22 crc kubenswrapper[4806]: E1204 05:26:22.028024 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c044b665-2475-438c-9616-f8856947509a" containerName="extract-utilities" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.028040 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c044b665-2475-438c-9616-f8856947509a" containerName="extract-utilities" Dec 04 05:26:22 crc kubenswrapper[4806]: E1204 05:26:22.028050 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5e571a-32a2-401e-9370-af762975718f" containerName="registry-server" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.028056 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5e571a-32a2-401e-9370-af762975718f" containerName="registry-server" Dec 04 05:26:22 crc kubenswrapper[4806]: E1204 05:26:22.028076 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5e571a-32a2-401e-9370-af762975718f" containerName="extract-utilities" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.028082 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5e571a-32a2-401e-9370-af762975718f" containerName="extract-utilities" Dec 04 05:26:22 crc kubenswrapper[4806]: E1204 05:26:22.028109 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c044b665-2475-438c-9616-f8856947509a" containerName="extract-content" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.028116 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c044b665-2475-438c-9616-f8856947509a" containerName="extract-content" Dec 04 05:26:22 crc kubenswrapper[4806]: E1204 05:26:22.028128 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5e571a-32a2-401e-9370-af762975718f" containerName="extract-content" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.028133 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5e571a-32a2-401e-9370-af762975718f" containerName="extract-content" Dec 04 05:26:22 crc kubenswrapper[4806]: E1204 05:26:22.028151 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c044b665-2475-438c-9616-f8856947509a" containerName="registry-server" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.028156 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="c044b665-2475-438c-9616-f8856947509a" containerName="registry-server" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.028338 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="c044b665-2475-438c-9616-f8856947509a" containerName="registry-server" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.028352 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c5e571a-32a2-401e-9370-af762975718f" containerName="registry-server" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.029681 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.044086 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tgg8"] Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.124126 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/214ee953-0140-4305-8850-c1780745d735-catalog-content\") pod \"redhat-marketplace-7tgg8\" (UID: \"214ee953-0140-4305-8850-c1780745d735\") " pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.124188 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft9sz\" (UniqueName: \"kubernetes.io/projected/214ee953-0140-4305-8850-c1780745d735-kube-api-access-ft9sz\") pod \"redhat-marketplace-7tgg8\" (UID: \"214ee953-0140-4305-8850-c1780745d735\") " pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.124334 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/214ee953-0140-4305-8850-c1780745d735-utilities\") pod \"redhat-marketplace-7tgg8\" (UID: \"214ee953-0140-4305-8850-c1780745d735\") " pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.225763 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/214ee953-0140-4305-8850-c1780745d735-utilities\") pod \"redhat-marketplace-7tgg8\" (UID: \"214ee953-0140-4305-8850-c1780745d735\") " pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.225896 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/214ee953-0140-4305-8850-c1780745d735-catalog-content\") pod \"redhat-marketplace-7tgg8\" (UID: \"214ee953-0140-4305-8850-c1780745d735\") " pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.225949 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft9sz\" (UniqueName: \"kubernetes.io/projected/214ee953-0140-4305-8850-c1780745d735-kube-api-access-ft9sz\") pod \"redhat-marketplace-7tgg8\" (UID: \"214ee953-0140-4305-8850-c1780745d735\") " pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.226301 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/214ee953-0140-4305-8850-c1780745d735-utilities\") pod \"redhat-marketplace-7tgg8\" (UID: \"214ee953-0140-4305-8850-c1780745d735\") " pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.226561 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/214ee953-0140-4305-8850-c1780745d735-catalog-content\") pod \"redhat-marketplace-7tgg8\" (UID: \"214ee953-0140-4305-8850-c1780745d735\") " pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.246496 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft9sz\" (UniqueName: \"kubernetes.io/projected/214ee953-0140-4305-8850-c1780745d735-kube-api-access-ft9sz\") pod \"redhat-marketplace-7tgg8\" (UID: \"214ee953-0140-4305-8850-c1780745d735\") " pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.395119 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:22 crc kubenswrapper[4806]: I1204 05:26:22.986511 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tgg8"] Dec 04 05:26:23 crc kubenswrapper[4806]: I1204 05:26:23.287259 4806 generic.go:334] "Generic (PLEG): container finished" podID="214ee953-0140-4305-8850-c1780745d735" containerID="a5d3209cf62fac8207d2198a05b658e990493bcb230f24c7c2ac0ff2ffb5a4b2" exitCode=0 Dec 04 05:26:23 crc kubenswrapper[4806]: I1204 05:26:23.287344 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tgg8" event={"ID":"214ee953-0140-4305-8850-c1780745d735","Type":"ContainerDied","Data":"a5d3209cf62fac8207d2198a05b658e990493bcb230f24c7c2ac0ff2ffb5a4b2"} Dec 04 05:26:23 crc kubenswrapper[4806]: I1204 05:26:23.287570 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tgg8" event={"ID":"214ee953-0140-4305-8850-c1780745d735","Type":"ContainerStarted","Data":"ccf3d5840f112e11eeb466a1fadba0c984bfe67c7a7417ac27b40a329f20c2dd"} Dec 04 05:26:24 crc kubenswrapper[4806]: I1204 05:26:24.297263 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tgg8" event={"ID":"214ee953-0140-4305-8850-c1780745d735","Type":"ContainerStarted","Data":"c22028dc3a3600569d3cbdd6e45a1894bd3904d87321fae762e5038c5b0fbaf3"} Dec 04 05:26:25 crc kubenswrapper[4806]: I1204 05:26:25.306385 4806 generic.go:334] "Generic (PLEG): container finished" podID="214ee953-0140-4305-8850-c1780745d735" containerID="c22028dc3a3600569d3cbdd6e45a1894bd3904d87321fae762e5038c5b0fbaf3" exitCode=0 Dec 04 05:26:25 crc kubenswrapper[4806]: I1204 05:26:25.306477 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tgg8" event={"ID":"214ee953-0140-4305-8850-c1780745d735","Type":"ContainerDied","Data":"c22028dc3a3600569d3cbdd6e45a1894bd3904d87321fae762e5038c5b0fbaf3"} Dec 04 05:26:26 crc kubenswrapper[4806]: I1204 05:26:26.316770 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tgg8" event={"ID":"214ee953-0140-4305-8850-c1780745d735","Type":"ContainerStarted","Data":"7f8d8a69f0a6c5c40ddf1132d5ac0d3254033d12f4e7d4381c3076780151cf2e"} Dec 04 05:26:26 crc kubenswrapper[4806]: I1204 05:26:26.343461 4806 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7tgg8" podStartSLOduration=2.935115409 podStartE2EDuration="5.343442058s" podCreationTimestamp="2025-12-04 05:26:21 +0000 UTC" firstStartedPulling="2025-12-04 05:26:23.290357081 +0000 UTC m=+5498.148870009" lastFinishedPulling="2025-12-04 05:26:25.69868373 +0000 UTC m=+5500.557196658" observedRunningTime="2025-12-04 05:26:26.339989128 +0000 UTC m=+5501.198502056" watchObservedRunningTime="2025-12-04 05:26:26.343442058 +0000 UTC m=+5501.201954986" Dec 04 05:26:32 crc kubenswrapper[4806]: I1204 05:26:32.395946 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:32 crc kubenswrapper[4806]: I1204 05:26:32.396480 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:32 crc kubenswrapper[4806]: I1204 05:26:32.446359 4806 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:33 crc kubenswrapper[4806]: I1204 05:26:33.125311 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-qq2x8_d80ba885-bdec-472c-b9eb-23928eeb6ab1/kube-rbac-proxy/0.log" Dec 04 05:26:33 crc kubenswrapper[4806]: I1204 05:26:33.385619 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-qq2x8_d80ba885-bdec-472c-b9eb-23928eeb6ab1/controller/0.log" Dec 04 05:26:33 crc kubenswrapper[4806]: I1204 05:26:33.425335 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-frr-files/0.log" Dec 04 05:26:33 crc kubenswrapper[4806]: I1204 05:26:33.436746 4806 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:33 crc kubenswrapper[4806]: I1204 05:26:33.481509 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tgg8"] Dec 04 05:26:33 crc kubenswrapper[4806]: I1204 05:26:33.588521 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-frr-files/0.log" Dec 04 05:26:33 crc kubenswrapper[4806]: I1204 05:26:33.635712 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-reloader/0.log" Dec 04 05:26:33 crc kubenswrapper[4806]: I1204 05:26:33.642892 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-reloader/0.log" Dec 04 05:26:33 crc kubenswrapper[4806]: I1204 05:26:33.703359 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-metrics/0.log" Dec 04 05:26:33 crc kubenswrapper[4806]: I1204 05:26:33.859428 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-frr-files/0.log" Dec 04 05:26:33 crc kubenswrapper[4806]: I1204 05:26:33.914768 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-metrics/0.log" Dec 04 05:26:33 crc kubenswrapper[4806]: I1204 05:26:33.952092 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-metrics/0.log" Dec 04 05:26:33 crc kubenswrapper[4806]: I1204 05:26:33.961383 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-reloader/0.log" Dec 04 05:26:34 crc kubenswrapper[4806]: I1204 05:26:34.132113 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-frr-files/0.log" Dec 04 05:26:34 crc kubenswrapper[4806]: I1204 05:26:34.134277 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-metrics/0.log" Dec 04 05:26:34 crc kubenswrapper[4806]: I1204 05:26:34.197752 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/cp-reloader/0.log" Dec 04 05:26:34 crc kubenswrapper[4806]: I1204 05:26:34.208917 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/controller/0.log" Dec 04 05:26:34 crc kubenswrapper[4806]: I1204 05:26:34.368804 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/frr-metrics/0.log" Dec 04 05:26:34 crc kubenswrapper[4806]: I1204 05:26:34.405450 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/kube-rbac-proxy/0.log" Dec 04 05:26:34 crc kubenswrapper[4806]: I1204 05:26:34.530855 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/kube-rbac-proxy-frr/0.log" Dec 04 05:26:34 crc kubenswrapper[4806]: I1204 05:26:34.639026 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/reloader/0.log" Dec 04 05:26:34 crc kubenswrapper[4806]: I1204 05:26:34.840298 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-wh624_fac17a35-6b82-4397-855c-a9d2b27bcca6/frr-k8s-webhook-server/0.log" Dec 04 05:26:35 crc kubenswrapper[4806]: I1204 05:26:35.020912 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5b99fcbdc-n6szh_09d57bc9-a5d5-4848-9cbd-12ce41ca2b78/manager/0.log" Dec 04 05:26:35 crc kubenswrapper[4806]: I1204 05:26:35.216065 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-58dfcfbb98-7prrp_134cc417-ef0e-4cb7-90e5-5a8f55466f6a/webhook-server/0.log" Dec 04 05:26:35 crc kubenswrapper[4806]: I1204 05:26:35.394450 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7tgg8" podUID="214ee953-0140-4305-8850-c1780745d735" containerName="registry-server" containerID="cri-o://7f8d8a69f0a6c5c40ddf1132d5ac0d3254033d12f4e7d4381c3076780151cf2e" gracePeriod=2 Dec 04 05:26:35 crc kubenswrapper[4806]: I1204 05:26:35.523003 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-g44sf_e653dfc5-42b0-4ae3-a622-b64b556d3ccc/kube-rbac-proxy/0.log" Dec 04 05:26:35 crc kubenswrapper[4806]: I1204 05:26:35.918291 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.009752 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r5wz_cb278465-e6d8-48b2-bc17-861968bba45e/frr/0.log" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.069257 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ft9sz\" (UniqueName: \"kubernetes.io/projected/214ee953-0140-4305-8850-c1780745d735-kube-api-access-ft9sz\") pod \"214ee953-0140-4305-8850-c1780745d735\" (UID: \"214ee953-0140-4305-8850-c1780745d735\") " Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.070062 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/214ee953-0140-4305-8850-c1780745d735-utilities\") pod \"214ee953-0140-4305-8850-c1780745d735\" (UID: \"214ee953-0140-4305-8850-c1780745d735\") " Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.070147 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/214ee953-0140-4305-8850-c1780745d735-catalog-content\") pod \"214ee953-0140-4305-8850-c1780745d735\" (UID: \"214ee953-0140-4305-8850-c1780745d735\") " Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.071853 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/214ee953-0140-4305-8850-c1780745d735-utilities" (OuterVolumeSpecName: "utilities") pod "214ee953-0140-4305-8850-c1780745d735" (UID: "214ee953-0140-4305-8850-c1780745d735"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.079688 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-g44sf_e653dfc5-42b0-4ae3-a622-b64b556d3ccc/speaker/0.log" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.083147 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/214ee953-0140-4305-8850-c1780745d735-kube-api-access-ft9sz" (OuterVolumeSpecName: "kube-api-access-ft9sz") pod "214ee953-0140-4305-8850-c1780745d735" (UID: "214ee953-0140-4305-8850-c1780745d735"). InnerVolumeSpecName "kube-api-access-ft9sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.100586 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/214ee953-0140-4305-8850-c1780745d735-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "214ee953-0140-4305-8850-c1780745d735" (UID: "214ee953-0140-4305-8850-c1780745d735"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.171952 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ft9sz\" (UniqueName: \"kubernetes.io/projected/214ee953-0140-4305-8850-c1780745d735-kube-api-access-ft9sz\") on node \"crc\" DevicePath \"\"" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.171996 4806 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/214ee953-0140-4305-8850-c1780745d735-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.172009 4806 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/214ee953-0140-4305-8850-c1780745d735-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.406594 4806 generic.go:334] "Generic (PLEG): container finished" podID="214ee953-0140-4305-8850-c1780745d735" containerID="7f8d8a69f0a6c5c40ddf1132d5ac0d3254033d12f4e7d4381c3076780151cf2e" exitCode=0 Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.407047 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tgg8" event={"ID":"214ee953-0140-4305-8850-c1780745d735","Type":"ContainerDied","Data":"7f8d8a69f0a6c5c40ddf1132d5ac0d3254033d12f4e7d4381c3076780151cf2e"} Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.407180 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tgg8" event={"ID":"214ee953-0140-4305-8850-c1780745d735","Type":"ContainerDied","Data":"ccf3d5840f112e11eeb466a1fadba0c984bfe67c7a7417ac27b40a329f20c2dd"} Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.407286 4806 scope.go:117] "RemoveContainer" containerID="7f8d8a69f0a6c5c40ddf1132d5ac0d3254033d12f4e7d4381c3076780151cf2e" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.407570 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7tgg8" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.433901 4806 scope.go:117] "RemoveContainer" containerID="c22028dc3a3600569d3cbdd6e45a1894bd3904d87321fae762e5038c5b0fbaf3" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.452887 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tgg8"] Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.465825 4806 scope.go:117] "RemoveContainer" containerID="a5d3209cf62fac8207d2198a05b658e990493bcb230f24c7c2ac0ff2ffb5a4b2" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.473202 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tgg8"] Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.521072 4806 scope.go:117] "RemoveContainer" containerID="7f8d8a69f0a6c5c40ddf1132d5ac0d3254033d12f4e7d4381c3076780151cf2e" Dec 04 05:26:36 crc kubenswrapper[4806]: E1204 05:26:36.521536 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f8d8a69f0a6c5c40ddf1132d5ac0d3254033d12f4e7d4381c3076780151cf2e\": container with ID starting with 7f8d8a69f0a6c5c40ddf1132d5ac0d3254033d12f4e7d4381c3076780151cf2e not found: ID does not exist" containerID="7f8d8a69f0a6c5c40ddf1132d5ac0d3254033d12f4e7d4381c3076780151cf2e" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.521576 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f8d8a69f0a6c5c40ddf1132d5ac0d3254033d12f4e7d4381c3076780151cf2e"} err="failed to get container status \"7f8d8a69f0a6c5c40ddf1132d5ac0d3254033d12f4e7d4381c3076780151cf2e\": rpc error: code = NotFound desc = could not find container \"7f8d8a69f0a6c5c40ddf1132d5ac0d3254033d12f4e7d4381c3076780151cf2e\": container with ID starting with 7f8d8a69f0a6c5c40ddf1132d5ac0d3254033d12f4e7d4381c3076780151cf2e not found: ID does not exist" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.521603 4806 scope.go:117] "RemoveContainer" containerID="c22028dc3a3600569d3cbdd6e45a1894bd3904d87321fae762e5038c5b0fbaf3" Dec 04 05:26:36 crc kubenswrapper[4806]: E1204 05:26:36.521878 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c22028dc3a3600569d3cbdd6e45a1894bd3904d87321fae762e5038c5b0fbaf3\": container with ID starting with c22028dc3a3600569d3cbdd6e45a1894bd3904d87321fae762e5038c5b0fbaf3 not found: ID does not exist" containerID="c22028dc3a3600569d3cbdd6e45a1894bd3904d87321fae762e5038c5b0fbaf3" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.521914 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c22028dc3a3600569d3cbdd6e45a1894bd3904d87321fae762e5038c5b0fbaf3"} err="failed to get container status \"c22028dc3a3600569d3cbdd6e45a1894bd3904d87321fae762e5038c5b0fbaf3\": rpc error: code = NotFound desc = could not find container \"c22028dc3a3600569d3cbdd6e45a1894bd3904d87321fae762e5038c5b0fbaf3\": container with ID starting with c22028dc3a3600569d3cbdd6e45a1894bd3904d87321fae762e5038c5b0fbaf3 not found: ID does not exist" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.521998 4806 scope.go:117] "RemoveContainer" containerID="a5d3209cf62fac8207d2198a05b658e990493bcb230f24c7c2ac0ff2ffb5a4b2" Dec 04 05:26:36 crc kubenswrapper[4806]: E1204 05:26:36.522242 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5d3209cf62fac8207d2198a05b658e990493bcb230f24c7c2ac0ff2ffb5a4b2\": container with ID starting with a5d3209cf62fac8207d2198a05b658e990493bcb230f24c7c2ac0ff2ffb5a4b2 not found: ID does not exist" containerID="a5d3209cf62fac8207d2198a05b658e990493bcb230f24c7c2ac0ff2ffb5a4b2" Dec 04 05:26:36 crc kubenswrapper[4806]: I1204 05:26:36.522269 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5d3209cf62fac8207d2198a05b658e990493bcb230f24c7c2ac0ff2ffb5a4b2"} err="failed to get container status \"a5d3209cf62fac8207d2198a05b658e990493bcb230f24c7c2ac0ff2ffb5a4b2\": rpc error: code = NotFound desc = could not find container \"a5d3209cf62fac8207d2198a05b658e990493bcb230f24c7c2ac0ff2ffb5a4b2\": container with ID starting with a5d3209cf62fac8207d2198a05b658e990493bcb230f24c7c2ac0ff2ffb5a4b2 not found: ID does not exist" Dec 04 05:26:37 crc kubenswrapper[4806]: I1204 05:26:37.434319 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="214ee953-0140-4305-8850-c1780745d735" path="/var/lib/kubelet/pods/214ee953-0140-4305-8850-c1780745d735/volumes" Dec 04 05:26:49 crc kubenswrapper[4806]: I1204 05:26:49.677234 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b_8981d20c-af12-4497-abc8-f17930ec12c1/util/0.log" Dec 04 05:26:49 crc kubenswrapper[4806]: I1204 05:26:49.869918 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b_8981d20c-af12-4497-abc8-f17930ec12c1/util/0.log" Dec 04 05:26:49 crc kubenswrapper[4806]: I1204 05:26:49.889885 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b_8981d20c-af12-4497-abc8-f17930ec12c1/pull/0.log" Dec 04 05:26:49 crc kubenswrapper[4806]: I1204 05:26:49.915832 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b_8981d20c-af12-4497-abc8-f17930ec12c1/pull/0.log" Dec 04 05:26:50 crc kubenswrapper[4806]: I1204 05:26:50.083515 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b_8981d20c-af12-4497-abc8-f17930ec12c1/util/0.log" Dec 04 05:26:50 crc kubenswrapper[4806]: I1204 05:26:50.130312 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b_8981d20c-af12-4497-abc8-f17930ec12c1/pull/0.log" Dec 04 05:26:50 crc kubenswrapper[4806]: I1204 05:26:50.186783 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fpk89b_8981d20c-af12-4497-abc8-f17930ec12c1/extract/0.log" Dec 04 05:26:50 crc kubenswrapper[4806]: I1204 05:26:50.278202 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4_205b1465-25a7-43a5-9449-78baf9a87116/util/0.log" Dec 04 05:26:50 crc kubenswrapper[4806]: I1204 05:26:50.467295 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4_205b1465-25a7-43a5-9449-78baf9a87116/pull/0.log" Dec 04 05:26:50 crc kubenswrapper[4806]: I1204 05:26:50.480562 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4_205b1465-25a7-43a5-9449-78baf9a87116/util/0.log" Dec 04 05:26:50 crc kubenswrapper[4806]: I1204 05:26:50.515743 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4_205b1465-25a7-43a5-9449-78baf9a87116/pull/0.log" Dec 04 05:26:50 crc kubenswrapper[4806]: I1204 05:26:50.634590 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4_205b1465-25a7-43a5-9449-78baf9a87116/util/0.log" Dec 04 05:26:50 crc kubenswrapper[4806]: I1204 05:26:50.667004 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4_205b1465-25a7-43a5-9449-78baf9a87116/pull/0.log" Dec 04 05:26:50 crc kubenswrapper[4806]: I1204 05:26:50.683427 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f834p8z4_205b1465-25a7-43a5-9449-78baf9a87116/extract/0.log" Dec 04 05:26:50 crc kubenswrapper[4806]: I1204 05:26:50.856600 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mkm6b_c3bcad8b-9a77-4d44-9bde-c3890fe6427c/extract-utilities/0.log" Dec 04 05:26:51 crc kubenswrapper[4806]: I1204 05:26:51.032431 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mkm6b_c3bcad8b-9a77-4d44-9bde-c3890fe6427c/extract-content/0.log" Dec 04 05:26:51 crc kubenswrapper[4806]: I1204 05:26:51.041878 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mkm6b_c3bcad8b-9a77-4d44-9bde-c3890fe6427c/extract-utilities/0.log" Dec 04 05:26:51 crc kubenswrapper[4806]: I1204 05:26:51.042008 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mkm6b_c3bcad8b-9a77-4d44-9bde-c3890fe6427c/extract-content/0.log" Dec 04 05:26:51 crc kubenswrapper[4806]: I1204 05:26:51.221956 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mkm6b_c3bcad8b-9a77-4d44-9bde-c3890fe6427c/extract-content/0.log" Dec 04 05:26:51 crc kubenswrapper[4806]: I1204 05:26:51.227329 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mkm6b_c3bcad8b-9a77-4d44-9bde-c3890fe6427c/extract-utilities/0.log" Dec 04 05:26:51 crc kubenswrapper[4806]: I1204 05:26:51.553641 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2bzpg_0bd6d288-1d0e-49a2-b9de-5437c5d22356/extract-utilities/0.log" Dec 04 05:26:51 crc kubenswrapper[4806]: I1204 05:26:51.885355 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2bzpg_0bd6d288-1d0e-49a2-b9de-5437c5d22356/extract-utilities/0.log" Dec 04 05:26:51 crc kubenswrapper[4806]: I1204 05:26:51.907312 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2bzpg_0bd6d288-1d0e-49a2-b9de-5437c5d22356/extract-content/0.log" Dec 04 05:26:51 crc kubenswrapper[4806]: I1204 05:26:51.909770 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mkm6b_c3bcad8b-9a77-4d44-9bde-c3890fe6427c/registry-server/0.log" Dec 04 05:26:51 crc kubenswrapper[4806]: I1204 05:26:51.916161 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2bzpg_0bd6d288-1d0e-49a2-b9de-5437c5d22356/extract-content/0.log" Dec 04 05:26:52 crc kubenswrapper[4806]: I1204 05:26:52.059646 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2bzpg_0bd6d288-1d0e-49a2-b9de-5437c5d22356/extract-utilities/0.log" Dec 04 05:26:52 crc kubenswrapper[4806]: I1204 05:26:52.105630 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2bzpg_0bd6d288-1d0e-49a2-b9de-5437c5d22356/extract-content/0.log" Dec 04 05:26:52 crc kubenswrapper[4806]: I1204 05:26:52.320948 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9rvth_60ce8b7c-5cf6-452c-835e-9ab84a66aa65/marketplace-operator/2.log" Dec 04 05:26:52 crc kubenswrapper[4806]: I1204 05:26:52.448464 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9rvth_60ce8b7c-5cf6-452c-835e-9ab84a66aa65/marketplace-operator/3.log" Dec 04 05:26:52 crc kubenswrapper[4806]: I1204 05:26:52.618953 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kgr4l_7bf8469a-647b-4b61-ab6c-89bbcc436ab9/extract-utilities/0.log" Dec 04 05:26:52 crc kubenswrapper[4806]: I1204 05:26:52.829044 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2bzpg_0bd6d288-1d0e-49a2-b9de-5437c5d22356/registry-server/0.log" Dec 04 05:26:52 crc kubenswrapper[4806]: I1204 05:26:52.881585 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kgr4l_7bf8469a-647b-4b61-ab6c-89bbcc436ab9/extract-utilities/0.log" Dec 04 05:26:52 crc kubenswrapper[4806]: I1204 05:26:52.961386 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kgr4l_7bf8469a-647b-4b61-ab6c-89bbcc436ab9/extract-content/0.log" Dec 04 05:26:52 crc kubenswrapper[4806]: I1204 05:26:52.969292 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kgr4l_7bf8469a-647b-4b61-ab6c-89bbcc436ab9/extract-content/0.log" Dec 04 05:26:53 crc kubenswrapper[4806]: I1204 05:26:53.133446 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kgr4l_7bf8469a-647b-4b61-ab6c-89bbcc436ab9/extract-utilities/0.log" Dec 04 05:26:53 crc kubenswrapper[4806]: I1204 05:26:53.242346 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kgr4l_7bf8469a-647b-4b61-ab6c-89bbcc436ab9/extract-content/0.log" Dec 04 05:26:53 crc kubenswrapper[4806]: I1204 05:26:53.345103 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-kgr4l_7bf8469a-647b-4b61-ab6c-89bbcc436ab9/registry-server/0.log" Dec 04 05:26:53 crc kubenswrapper[4806]: I1204 05:26:53.436519 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrtjm_242a6296-78ef-431a-8c26-522e66e5296c/extract-utilities/0.log" Dec 04 05:26:53 crc kubenswrapper[4806]: I1204 05:26:53.607030 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrtjm_242a6296-78ef-431a-8c26-522e66e5296c/extract-content/0.log" Dec 04 05:26:53 crc kubenswrapper[4806]: I1204 05:26:53.630044 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrtjm_242a6296-78ef-431a-8c26-522e66e5296c/extract-utilities/0.log" Dec 04 05:26:53 crc kubenswrapper[4806]: I1204 05:26:53.640582 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrtjm_242a6296-78ef-431a-8c26-522e66e5296c/extract-content/0.log" Dec 04 05:26:53 crc kubenswrapper[4806]: I1204 05:26:53.783627 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrtjm_242a6296-78ef-431a-8c26-522e66e5296c/extract-content/0.log" Dec 04 05:26:53 crc kubenswrapper[4806]: I1204 05:26:53.804605 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrtjm_242a6296-78ef-431a-8c26-522e66e5296c/extract-utilities/0.log" Dec 04 05:26:54 crc kubenswrapper[4806]: I1204 05:26:54.406764 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-zrtjm_242a6296-78ef-431a-8c26-522e66e5296c/registry-server/0.log" Dec 04 05:27:27 crc kubenswrapper[4806]: I1204 05:27:27.047653 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:27:27 crc kubenswrapper[4806]: I1204 05:27:27.048156 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:27:57 crc kubenswrapper[4806]: I1204 05:27:57.047261 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:27:57 crc kubenswrapper[4806]: I1204 05:27:57.047653 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:28:27 crc kubenswrapper[4806]: I1204 05:28:27.047250 4806 patch_prober.go:28] interesting pod/machine-config-daemon-zpps8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 05:28:27 crc kubenswrapper[4806]: I1204 05:28:27.047685 4806 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 05:28:27 crc kubenswrapper[4806]: I1204 05:28:27.047722 4806 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" Dec 04 05:28:27 crc kubenswrapper[4806]: I1204 05:28:27.048399 4806 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19"} pod="openshift-machine-config-operator/machine-config-daemon-zpps8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 05:28:27 crc kubenswrapper[4806]: I1204 05:28:27.048457 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerName="machine-config-daemon" containerID="cri-o://cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" gracePeriod=600 Dec 04 05:28:27 crc kubenswrapper[4806]: E1204 05:28:27.165552 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:28:27 crc kubenswrapper[4806]: I1204 05:28:27.543342 4806 generic.go:334] "Generic (PLEG): container finished" podID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" exitCode=0 Dec 04 05:28:27 crc kubenswrapper[4806]: I1204 05:28:27.543384 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" event={"ID":"d0f9623b-44ce-43c2-b990-a75db2d17ff8","Type":"ContainerDied","Data":"cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19"} Dec 04 05:28:27 crc kubenswrapper[4806]: I1204 05:28:27.543414 4806 scope.go:117] "RemoveContainer" containerID="a025afa4e0f2bafc160c671d65ca45902522db3c999cd6d04dd15c40321dcf4a" Dec 04 05:28:27 crc kubenswrapper[4806]: I1204 05:28:27.544007 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:28:27 crc kubenswrapper[4806]: E1204 05:28:27.544225 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:28:38 crc kubenswrapper[4806]: I1204 05:28:38.423213 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:28:38 crc kubenswrapper[4806]: E1204 05:28:38.423970 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:28:51 crc kubenswrapper[4806]: I1204 05:28:51.424723 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:28:51 crc kubenswrapper[4806]: E1204 05:28:51.425468 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:29:06 crc kubenswrapper[4806]: I1204 05:29:06.424896 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:29:06 crc kubenswrapper[4806]: E1204 05:29:06.428414 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:29:09 crc kubenswrapper[4806]: I1204 05:29:09.018416 4806 generic.go:334] "Generic (PLEG): container finished" podID="ed7a3f69-8963-405c-b661-3763e87d2a69" containerID="cc336c30daa805af38a23f52cfb34297e2da0cc24d084abde460b4c4a02d7b50" exitCode=0 Dec 04 05:29:09 crc kubenswrapper[4806]: I1204 05:29:09.018525 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-dz69k/must-gather-zmxrl" event={"ID":"ed7a3f69-8963-405c-b661-3763e87d2a69","Type":"ContainerDied","Data":"cc336c30daa805af38a23f52cfb34297e2da0cc24d084abde460b4c4a02d7b50"} Dec 04 05:29:09 crc kubenswrapper[4806]: I1204 05:29:09.020565 4806 scope.go:117] "RemoveContainer" containerID="cc336c30daa805af38a23f52cfb34297e2da0cc24d084abde460b4c4a02d7b50" Dec 04 05:29:10 crc kubenswrapper[4806]: I1204 05:29:10.044068 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-dz69k_must-gather-zmxrl_ed7a3f69-8963-405c-b661-3763e87d2a69/gather/0.log" Dec 04 05:29:18 crc kubenswrapper[4806]: I1204 05:29:18.423872 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:29:18 crc kubenswrapper[4806]: E1204 05:29:18.425034 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:29:22 crc kubenswrapper[4806]: I1204 05:29:22.474338 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-dz69k/must-gather-zmxrl"] Dec 04 05:29:22 crc kubenswrapper[4806]: I1204 05:29:22.475223 4806 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-dz69k/must-gather-zmxrl" podUID="ed7a3f69-8963-405c-b661-3763e87d2a69" containerName="copy" containerID="cri-o://8cac059cac8ee864d72474a8e6d0307b6668bead6bdf640dfcfd5c0d86f58919" gracePeriod=2 Dec 04 05:29:22 crc kubenswrapper[4806]: I1204 05:29:22.487205 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-dz69k/must-gather-zmxrl"] Dec 04 05:29:22 crc kubenswrapper[4806]: I1204 05:29:22.938545 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-dz69k_must-gather-zmxrl_ed7a3f69-8963-405c-b661-3763e87d2a69/copy/0.log" Dec 04 05:29:22 crc kubenswrapper[4806]: I1204 05:29:22.939152 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dz69k/must-gather-zmxrl" Dec 04 05:29:23 crc kubenswrapper[4806]: I1204 05:29:23.078363 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ed7a3f69-8963-405c-b661-3763e87d2a69-must-gather-output\") pod \"ed7a3f69-8963-405c-b661-3763e87d2a69\" (UID: \"ed7a3f69-8963-405c-b661-3763e87d2a69\") " Dec 04 05:29:23 crc kubenswrapper[4806]: I1204 05:29:23.078648 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzllh\" (UniqueName: \"kubernetes.io/projected/ed7a3f69-8963-405c-b661-3763e87d2a69-kube-api-access-rzllh\") pod \"ed7a3f69-8963-405c-b661-3763e87d2a69\" (UID: \"ed7a3f69-8963-405c-b661-3763e87d2a69\") " Dec 04 05:29:23 crc kubenswrapper[4806]: I1204 05:29:23.087596 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed7a3f69-8963-405c-b661-3763e87d2a69-kube-api-access-rzllh" (OuterVolumeSpecName: "kube-api-access-rzllh") pod "ed7a3f69-8963-405c-b661-3763e87d2a69" (UID: "ed7a3f69-8963-405c-b661-3763e87d2a69"). InnerVolumeSpecName "kube-api-access-rzllh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:29:23 crc kubenswrapper[4806]: I1204 05:29:23.181778 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzllh\" (UniqueName: \"kubernetes.io/projected/ed7a3f69-8963-405c-b661-3763e87d2a69-kube-api-access-rzllh\") on node \"crc\" DevicePath \"\"" Dec 04 05:29:23 crc kubenswrapper[4806]: I1204 05:29:23.183189 4806 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-dz69k_must-gather-zmxrl_ed7a3f69-8963-405c-b661-3763e87d2a69/copy/0.log" Dec 04 05:29:23 crc kubenswrapper[4806]: I1204 05:29:23.183535 4806 generic.go:334] "Generic (PLEG): container finished" podID="ed7a3f69-8963-405c-b661-3763e87d2a69" containerID="8cac059cac8ee864d72474a8e6d0307b6668bead6bdf640dfcfd5c0d86f58919" exitCode=143 Dec 04 05:29:23 crc kubenswrapper[4806]: I1204 05:29:23.183578 4806 scope.go:117] "RemoveContainer" containerID="8cac059cac8ee864d72474a8e6d0307b6668bead6bdf640dfcfd5c0d86f58919" Dec 04 05:29:23 crc kubenswrapper[4806]: I1204 05:29:23.183680 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-dz69k/must-gather-zmxrl" Dec 04 05:29:23 crc kubenswrapper[4806]: I1204 05:29:23.241166 4806 scope.go:117] "RemoveContainer" containerID="cc336c30daa805af38a23f52cfb34297e2da0cc24d084abde460b4c4a02d7b50" Dec 04 05:29:23 crc kubenswrapper[4806]: I1204 05:29:23.265859 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed7a3f69-8963-405c-b661-3763e87d2a69-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "ed7a3f69-8963-405c-b661-3763e87d2a69" (UID: "ed7a3f69-8963-405c-b661-3763e87d2a69"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 05:29:23 crc kubenswrapper[4806]: I1204 05:29:23.283596 4806 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ed7a3f69-8963-405c-b661-3763e87d2a69-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 04 05:29:23 crc kubenswrapper[4806]: I1204 05:29:23.287715 4806 scope.go:117] "RemoveContainer" containerID="8cac059cac8ee864d72474a8e6d0307b6668bead6bdf640dfcfd5c0d86f58919" Dec 04 05:29:23 crc kubenswrapper[4806]: E1204 05:29:23.290096 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cac059cac8ee864d72474a8e6d0307b6668bead6bdf640dfcfd5c0d86f58919\": container with ID starting with 8cac059cac8ee864d72474a8e6d0307b6668bead6bdf640dfcfd5c0d86f58919 not found: ID does not exist" containerID="8cac059cac8ee864d72474a8e6d0307b6668bead6bdf640dfcfd5c0d86f58919" Dec 04 05:29:23 crc kubenswrapper[4806]: I1204 05:29:23.290143 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cac059cac8ee864d72474a8e6d0307b6668bead6bdf640dfcfd5c0d86f58919"} err="failed to get container status \"8cac059cac8ee864d72474a8e6d0307b6668bead6bdf640dfcfd5c0d86f58919\": rpc error: code = NotFound desc = could not find container \"8cac059cac8ee864d72474a8e6d0307b6668bead6bdf640dfcfd5c0d86f58919\": container with ID starting with 8cac059cac8ee864d72474a8e6d0307b6668bead6bdf640dfcfd5c0d86f58919 not found: ID does not exist" Dec 04 05:29:23 crc kubenswrapper[4806]: I1204 05:29:23.290173 4806 scope.go:117] "RemoveContainer" containerID="cc336c30daa805af38a23f52cfb34297e2da0cc24d084abde460b4c4a02d7b50" Dec 04 05:29:23 crc kubenswrapper[4806]: E1204 05:29:23.290674 4806 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc336c30daa805af38a23f52cfb34297e2da0cc24d084abde460b4c4a02d7b50\": container with ID starting with cc336c30daa805af38a23f52cfb34297e2da0cc24d084abde460b4c4a02d7b50 not found: ID does not exist" containerID="cc336c30daa805af38a23f52cfb34297e2da0cc24d084abde460b4c4a02d7b50" Dec 04 05:29:23 crc kubenswrapper[4806]: I1204 05:29:23.290706 4806 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc336c30daa805af38a23f52cfb34297e2da0cc24d084abde460b4c4a02d7b50"} err="failed to get container status \"cc336c30daa805af38a23f52cfb34297e2da0cc24d084abde460b4c4a02d7b50\": rpc error: code = NotFound desc = could not find container \"cc336c30daa805af38a23f52cfb34297e2da0cc24d084abde460b4c4a02d7b50\": container with ID starting with cc336c30daa805af38a23f52cfb34297e2da0cc24d084abde460b4c4a02d7b50 not found: ID does not exist" Dec 04 05:29:23 crc kubenswrapper[4806]: I1204 05:29:23.437135 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed7a3f69-8963-405c-b661-3763e87d2a69" path="/var/lib/kubelet/pods/ed7a3f69-8963-405c-b661-3763e87d2a69/volumes" Dec 04 05:29:29 crc kubenswrapper[4806]: I1204 05:29:29.428750 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:29:29 crc kubenswrapper[4806]: E1204 05:29:29.429576 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:29:40 crc kubenswrapper[4806]: I1204 05:29:40.424737 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:29:40 crc kubenswrapper[4806]: E1204 05:29:40.426173 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:29:52 crc kubenswrapper[4806]: I1204 05:29:52.423501 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:29:52 crc kubenswrapper[4806]: E1204 05:29:52.424859 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.158146 4806 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr"] Dec 04 05:30:00 crc kubenswrapper[4806]: E1204 05:30:00.163276 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="214ee953-0140-4305-8850-c1780745d735" containerName="extract-utilities" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.163306 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="214ee953-0140-4305-8850-c1780745d735" containerName="extract-utilities" Dec 04 05:30:00 crc kubenswrapper[4806]: E1204 05:30:00.163329 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed7a3f69-8963-405c-b661-3763e87d2a69" containerName="gather" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.163335 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed7a3f69-8963-405c-b661-3763e87d2a69" containerName="gather" Dec 04 05:30:00 crc kubenswrapper[4806]: E1204 05:30:00.163378 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="214ee953-0140-4305-8850-c1780745d735" containerName="registry-server" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.163385 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="214ee953-0140-4305-8850-c1780745d735" containerName="registry-server" Dec 04 05:30:00 crc kubenswrapper[4806]: E1204 05:30:00.163403 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed7a3f69-8963-405c-b661-3763e87d2a69" containerName="copy" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.163428 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed7a3f69-8963-405c-b661-3763e87d2a69" containerName="copy" Dec 04 05:30:00 crc kubenswrapper[4806]: E1204 05:30:00.163439 4806 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="214ee953-0140-4305-8850-c1780745d735" containerName="extract-content" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.163445 4806 state_mem.go:107] "Deleted CPUSet assignment" podUID="214ee953-0140-4305-8850-c1780745d735" containerName="extract-content" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.163697 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="214ee953-0140-4305-8850-c1780745d735" containerName="registry-server" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.163710 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed7a3f69-8963-405c-b661-3763e87d2a69" containerName="copy" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.163724 4806 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed7a3f69-8963-405c-b661-3763e87d2a69" containerName="gather" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.164413 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.171162 4806 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.172541 4806 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.175685 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr"] Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.245291 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59dffe21-b7b2-4016-a071-2cd8fc1516c0-secret-volume\") pod \"collect-profiles-29413770-vv8vr\" (UID: \"59dffe21-b7b2-4016-a071-2cd8fc1516c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.245365 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59dffe21-b7b2-4016-a071-2cd8fc1516c0-config-volume\") pod \"collect-profiles-29413770-vv8vr\" (UID: \"59dffe21-b7b2-4016-a071-2cd8fc1516c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.245633 4806 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6q6n\" (UniqueName: \"kubernetes.io/projected/59dffe21-b7b2-4016-a071-2cd8fc1516c0-kube-api-access-k6q6n\") pod \"collect-profiles-29413770-vv8vr\" (UID: \"59dffe21-b7b2-4016-a071-2cd8fc1516c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.348808 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59dffe21-b7b2-4016-a071-2cd8fc1516c0-secret-volume\") pod \"collect-profiles-29413770-vv8vr\" (UID: \"59dffe21-b7b2-4016-a071-2cd8fc1516c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.348900 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59dffe21-b7b2-4016-a071-2cd8fc1516c0-config-volume\") pod \"collect-profiles-29413770-vv8vr\" (UID: \"59dffe21-b7b2-4016-a071-2cd8fc1516c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.349092 4806 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6q6n\" (UniqueName: \"kubernetes.io/projected/59dffe21-b7b2-4016-a071-2cd8fc1516c0-kube-api-access-k6q6n\") pod \"collect-profiles-29413770-vv8vr\" (UID: \"59dffe21-b7b2-4016-a071-2cd8fc1516c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.349768 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59dffe21-b7b2-4016-a071-2cd8fc1516c0-config-volume\") pod \"collect-profiles-29413770-vv8vr\" (UID: \"59dffe21-b7b2-4016-a071-2cd8fc1516c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.359399 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59dffe21-b7b2-4016-a071-2cd8fc1516c0-secret-volume\") pod \"collect-profiles-29413770-vv8vr\" (UID: \"59dffe21-b7b2-4016-a071-2cd8fc1516c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.376167 4806 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6q6n\" (UniqueName: \"kubernetes.io/projected/59dffe21-b7b2-4016-a071-2cd8fc1516c0-kube-api-access-k6q6n\") pod \"collect-profiles-29413770-vv8vr\" (UID: \"59dffe21-b7b2-4016-a071-2cd8fc1516c0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr" Dec 04 05:30:00 crc kubenswrapper[4806]: I1204 05:30:00.514822 4806 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr" Dec 04 05:30:01 crc kubenswrapper[4806]: I1204 05:30:01.021305 4806 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr"] Dec 04 05:30:01 crc kubenswrapper[4806]: I1204 05:30:01.686439 4806 generic.go:334] "Generic (PLEG): container finished" podID="59dffe21-b7b2-4016-a071-2cd8fc1516c0" containerID="b178d6e0c0879d58ef3a4b500d40c8c358a96da10ef434bcb96e2ccd116bf980" exitCode=0 Dec 04 05:30:01 crc kubenswrapper[4806]: I1204 05:30:01.686508 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr" event={"ID":"59dffe21-b7b2-4016-a071-2cd8fc1516c0","Type":"ContainerDied","Data":"b178d6e0c0879d58ef3a4b500d40c8c358a96da10ef434bcb96e2ccd116bf980"} Dec 04 05:30:01 crc kubenswrapper[4806]: I1204 05:30:01.686675 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr" event={"ID":"59dffe21-b7b2-4016-a071-2cd8fc1516c0","Type":"ContainerStarted","Data":"fbdcb35af926f45f8627ee64530eea76e5a58a315b0740dc1f939aee9f3eef4c"} Dec 04 05:30:03 crc kubenswrapper[4806]: I1204 05:30:03.059958 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr" Dec 04 05:30:03 crc kubenswrapper[4806]: I1204 05:30:03.211352 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59dffe21-b7b2-4016-a071-2cd8fc1516c0-config-volume\") pod \"59dffe21-b7b2-4016-a071-2cd8fc1516c0\" (UID: \"59dffe21-b7b2-4016-a071-2cd8fc1516c0\") " Dec 04 05:30:03 crc kubenswrapper[4806]: I1204 05:30:03.211418 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6q6n\" (UniqueName: \"kubernetes.io/projected/59dffe21-b7b2-4016-a071-2cd8fc1516c0-kube-api-access-k6q6n\") pod \"59dffe21-b7b2-4016-a071-2cd8fc1516c0\" (UID: \"59dffe21-b7b2-4016-a071-2cd8fc1516c0\") " Dec 04 05:30:03 crc kubenswrapper[4806]: I1204 05:30:03.211515 4806 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59dffe21-b7b2-4016-a071-2cd8fc1516c0-secret-volume\") pod \"59dffe21-b7b2-4016-a071-2cd8fc1516c0\" (UID: \"59dffe21-b7b2-4016-a071-2cd8fc1516c0\") " Dec 04 05:30:03 crc kubenswrapper[4806]: I1204 05:30:03.212423 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59dffe21-b7b2-4016-a071-2cd8fc1516c0-config-volume" (OuterVolumeSpecName: "config-volume") pod "59dffe21-b7b2-4016-a071-2cd8fc1516c0" (UID: "59dffe21-b7b2-4016-a071-2cd8fc1516c0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 05:30:03 crc kubenswrapper[4806]: I1204 05:30:03.220968 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59dffe21-b7b2-4016-a071-2cd8fc1516c0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "59dffe21-b7b2-4016-a071-2cd8fc1516c0" (UID: "59dffe21-b7b2-4016-a071-2cd8fc1516c0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 05:30:03 crc kubenswrapper[4806]: I1204 05:30:03.221097 4806 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59dffe21-b7b2-4016-a071-2cd8fc1516c0-kube-api-access-k6q6n" (OuterVolumeSpecName: "kube-api-access-k6q6n") pod "59dffe21-b7b2-4016-a071-2cd8fc1516c0" (UID: "59dffe21-b7b2-4016-a071-2cd8fc1516c0"). InnerVolumeSpecName "kube-api-access-k6q6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 05:30:03 crc kubenswrapper[4806]: I1204 05:30:03.314282 4806 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6q6n\" (UniqueName: \"kubernetes.io/projected/59dffe21-b7b2-4016-a071-2cd8fc1516c0-kube-api-access-k6q6n\") on node \"crc\" DevicePath \"\"" Dec 04 05:30:03 crc kubenswrapper[4806]: I1204 05:30:03.314319 4806 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59dffe21-b7b2-4016-a071-2cd8fc1516c0-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 05:30:03 crc kubenswrapper[4806]: I1204 05:30:03.314332 4806 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59dffe21-b7b2-4016-a071-2cd8fc1516c0-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 05:30:03 crc kubenswrapper[4806]: I1204 05:30:03.697520 4806 scope.go:117] "RemoveContainer" containerID="243d8cffc55529b9a7c58e34203ce3ab6b166fa34cf27c329467655cafd3762c" Dec 04 05:30:03 crc kubenswrapper[4806]: I1204 05:30:03.713131 4806 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr" event={"ID":"59dffe21-b7b2-4016-a071-2cd8fc1516c0","Type":"ContainerDied","Data":"fbdcb35af926f45f8627ee64530eea76e5a58a315b0740dc1f939aee9f3eef4c"} Dec 04 05:30:03 crc kubenswrapper[4806]: I1204 05:30:03.713234 4806 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbdcb35af926f45f8627ee64530eea76e5a58a315b0740dc1f939aee9f3eef4c" Dec 04 05:30:03 crc kubenswrapper[4806]: I1204 05:30:03.713326 4806 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413770-vv8vr" Dec 04 05:30:04 crc kubenswrapper[4806]: I1204 05:30:04.175447 4806 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll"] Dec 04 05:30:04 crc kubenswrapper[4806]: I1204 05:30:04.186684 4806 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413725-4vnll"] Dec 04 05:30:04 crc kubenswrapper[4806]: I1204 05:30:04.424266 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:30:04 crc kubenswrapper[4806]: E1204 05:30:04.424725 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:30:05 crc kubenswrapper[4806]: I1204 05:30:05.456442 4806 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679" path="/var/lib/kubelet/pods/c9c4bf4a-1e2b-4d69-85d7-d0aa4f9bf679/volumes" Dec 04 05:30:17 crc kubenswrapper[4806]: I1204 05:30:17.424327 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:30:17 crc kubenswrapper[4806]: E1204 05:30:17.425216 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:30:32 crc kubenswrapper[4806]: I1204 05:30:32.423836 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:30:32 crc kubenswrapper[4806]: E1204 05:30:32.424757 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:30:46 crc kubenswrapper[4806]: I1204 05:30:46.423856 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:30:46 crc kubenswrapper[4806]: E1204 05:30:46.425467 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:30:57 crc kubenswrapper[4806]: I1204 05:30:57.423785 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:30:57 crc kubenswrapper[4806]: E1204 05:30:57.425004 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:31:03 crc kubenswrapper[4806]: I1204 05:31:03.791273 4806 scope.go:117] "RemoveContainer" containerID="28094f5cafb39713eed10d3b7ccffa5f7e4a97fb80832b06fd4c4c92e2759490" Dec 04 05:31:12 crc kubenswrapper[4806]: I1204 05:31:12.424108 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:31:12 crc kubenswrapper[4806]: E1204 05:31:12.425062 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:31:26 crc kubenswrapper[4806]: I1204 05:31:26.424196 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:31:26 crc kubenswrapper[4806]: E1204 05:31:26.425484 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:31:41 crc kubenswrapper[4806]: I1204 05:31:41.428047 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:31:41 crc kubenswrapper[4806]: E1204 05:31:41.428735 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:31:54 crc kubenswrapper[4806]: I1204 05:31:54.424273 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:31:54 crc kubenswrapper[4806]: E1204 05:31:54.425065 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:32:09 crc kubenswrapper[4806]: I1204 05:32:09.424794 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:32:09 crc kubenswrapper[4806]: E1204 05:32:09.426160 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:32:21 crc kubenswrapper[4806]: I1204 05:32:21.428056 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:32:21 crc kubenswrapper[4806]: E1204 05:32:21.429196 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:32:32 crc kubenswrapper[4806]: I1204 05:32:32.424337 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:32:32 crc kubenswrapper[4806]: E1204 05:32:32.425319 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:32:45 crc kubenswrapper[4806]: I1204 05:32:45.431199 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:32:45 crc kubenswrapper[4806]: E1204 05:32:45.431901 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:32:56 crc kubenswrapper[4806]: I1204 05:32:56.424197 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:32:56 crc kubenswrapper[4806]: E1204 05:32:56.424846 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" Dec 04 05:33:10 crc kubenswrapper[4806]: I1204 05:33:10.423557 4806 scope.go:117] "RemoveContainer" containerID="cbccec69b713d1ac9aa1526ceb8cbc7365d890b5055c339d6dbc3e153bcc6a19" Dec 04 05:33:10 crc kubenswrapper[4806]: E1204 05:33:10.424361 4806 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zpps8_openshift-machine-config-operator(d0f9623b-44ce-43c2-b990-a75db2d17ff8)\"" pod="openshift-machine-config-operator/machine-config-daemon-zpps8" podUID="d0f9623b-44ce-43c2-b990-a75db2d17ff8" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114216445024450 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114216446017366 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114202406016500 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114202406015450 5ustar corecore